var/home/core/zuul-output/0000755000175000017500000000000015154737115014536 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015154750343015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000307654615154750260020300 0ustar corecoreгikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ}3I_翪|mvſFެxۻf+ovpZjlC4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;F}Zl8T*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}+7 s'ϨF&%8'# $9b"r>B)GF%\bi/ Ff/Bp 4YH~BŊ6EZ|^߸3%L[EC 7gg/碓@e=Vn)h\\lwCzDiQJxTsL] ,=M`nͷ~Vܯ5a|X&pNz7l9HGAr Mme)M,O!Xa~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ]rwbXh)g''H_`!GKF5/O]Zڢ>:O񨡺ePӋ&56zGnL!?lJJYq=Wo/"IyQ4\:y|6h6dQX0>HTG5QOuxMe 1׶/5άRIo>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'P'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJBR_v'5n]FhNU˿oۂ6C9C7sn,kje*;iΓA7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amKGm+`VLJsC>?5rk{-3Ss`y_C}Q v,{*)ߎ% qƦat:D=uNvdߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B%C6Θ%|5u=kkN2{'FEc* A>{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t!X򴱞_aM:E.Qg1DllЊE҉L ehJx{̗Uɾ?si&2"C]u$.`mjmƒVe9f6NŐsLu6fe wkىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0!i(`Z;TyֻΗ|ִ0-6dAC5t[OM91c:VJR9&ksvJ;0ɝ$krogB= FYtЩOte=?>T&O{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSVO$.KMb.:DK>WtWǭKv4@Va3"a`R@gbu%_J5Ґ 3?lm$K/$s_. WM]̍"W%`lO2-"ew@E=! I,($F{ձ7*Oy 6EK( EF #31J8mN .TTF9㕴/5~RxCe,&v3,JE- ZF5%Da,Gܠ*qI@qlG6s푻jÝ$ >8ȕ$eZ1j[h0SH,qf<"${/ksBK}xnwDb%M6:K<~̓9*u᛹Q{FЖt~6S#G1(zr6<ߜ!?U\(0EmG4 4c~J~]ps/9܎ms4gZY-07`-Id,9õ԰t+-b[uemNi_󈛥^g+!SKq<>78NBx;c4<ニ)H .Pd^cR^p_G+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YLz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^oHmmU̸nG pHKZ{{Qo}i¿Xc\]e1e,5`te.5Hhao<[50wMUF􀍠PV?Yg"ź)\3mf|ܔMUiU|Ym! #'ukMmQ9Blm]TO1ba.XW x6ܠ9[v35H;-]Um4mMrW-k#~fؤϋu_j*^Wj^qM `-Pk.@%=X#|ۡb1lKcj$׋bKv[~"N jS4HOkeF3LPyi︅iWk! cAnxu6<7cp?WN $?X3l(?  'Z! ,Z.maO_Bk/m~ޖ(<qRfR"Au\PmLZ"twpuJ` mvf+T!6Ѓjw1ncuwo':o gSPC=]U҅yY9 &K<-na'Xk,P4+`Þ/lX/bjFO.= w ?>ȑ3n߿z,t s5Z/ Clo-` z?a~b mzkC zFȏ>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwoj-25Hݻ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@T%W?%TbzK-6cb:XeGL`'žeVVޖ~;BLv[n|viPjbMeO?!hEfޮ])4 ?KN1o<]0Bg9lldXuT ʑ!Iu2ʌnB5*<^I^~G;Ja߄bHȌsK+D"̽E/"Icƀsu0,gy(&TI{ U܋N5 l͖h"褁lm *#n/Q!m b0X3i)\IN˭% Y&cKoG w 9pM^WϋQf7s#bd+SDL ,FZ<1Kx&C!{P|Ռr,* ] O;*X]Eg,5,ouZm8pnglVj!p2֬uT[QyB402|2d5K: `Bcz|Rxxl3{c` 1nhJzQHv?hbºܞz=73qSO0}Dc D]ͺjgw07'㤸z YJ\Hb9Ɖ„2Hi{(2HFE?*w*hy4ޙM^٫wF(p]EwQzr*! 5F XrO7E[!gJ^.a&HߣaaQÝ$_vyz4}0!yܒ栒޹a% Ŋ X!cJ!A\ ?E\R1 q/rJjd A4y4c+bQ̘TT!kw/nb͵FcRG0xeO sw5TV12R7<OG5cjShGg/5TbW > ]~Wޠ9dNiee$V[\[Qp-&u~a+3~;xUFFW>'ǣC~방u)т48ZdH;j a]`bGԹ#qiP(yڤ~dO@wA[Vz/$NW\F?H4kX6)F*1*(eJAaݡ krqB}q^fn 8y7P  GRޠkQn>eqQntq"Occ°NRjg#qSn02DŔw:ؽ 5l)Fa/TTmCԤ{"9b{ywSXE*m#3U ùRIvޏrJ`k|wJKH:O*OKy`( ݢe*{ ua ȻݔhvOkU~OǠI/aǕ-JMX _.6KsjA Qsmd  O#F.Uf28ZAgy>y,d$C?v01q5e.Um>]RLa&r?+@6k&#l)I5_> ` D s5npo}/ؙq #a2V?X~.4O/'|/_|&q̑0dd4>vk 60D _o~[Sw3ckpkpLNa ^j 5*<&}kˢmqvۗj=<Tr=[ a^؃ È(<^=xZb [_tܡ&yЋ{ Sym^?̑sU~' Ԓ f\itu)b>5X -$wMm[eG`̵E$uLrk-$_{$# $B*hN/ٟPE[/Y5d{zrBܖ6Hlc "mKv~[uLU4lZ;xEN'oI㤛rP*jC# 6@dmHg1$ʇȠh#CBΤ{sTQ{%w)7@y1K^ ].Y$46[B-3%OONw8d`Q4d$x0t8@t]y1T\YAidtxBG:pɨyeNg4n]M؞ e}Wn6׳i~'ہZ*FU{fXڃP'Hl4 ,ŸqMHDCYZz Qnz܁$Jp04ȴIL΃.0FiO-qy)i_TA|S2G4miBȨHM(2hys|F 94 DNlϒòκ-q|xC ,gKDzHR%t+E/wd#礱ºȄWEz o\JξB.wLKZ39(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A7? /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"VS^΄t*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#;N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{W+ߑ?X2hS4-=^YgpUHެbZ!y!ul@ڼ63" ۩:6=TZõ$E,ϓRV|G&$rr;J TtIHFE=RȬ]P pLm|?$%>Eü%mWO[>Xmw,*9.[G n >X8Ī;xW%dT:`ٓ~:QO,}j6j!yڦʲT:Pqҋh] H+&=>g| Z;D8ܶb:! Å{2:+au 6:!fF+0#+̬NY"!6a7#񕪰%:r|o5Znڧs?si/W qEU馥˟^_޶oڷOj'?nc]Rn\t3^邳塨Lɏ"8k8M~?M}OAH$77f|lgn I%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|'$0&*m.)HzzBvU0h}~5[Z!]nlnݔn,?WTm>C9O n6oHNe">0]8@*0)QsUN8t^N+mXU q2EDö0^R) hCt{d}ܜFnԴ.2O⠪R/r| w,?VMqܙ7'qpUۚ5Tnj ۝jlN$q:w$U>tL)NC*<` `)ĉJآS2 z]GQ)Bی:D`W&jDk\XDy&?Y\9Ȣ{:${1`+iur?qzJJaQ#-n`-$fhnqgTĔO5 ꐌSYXzv9[ezksA`<dkON৯s|&*pNaJه5B5H:W2% `6MRR'xZtfC$1aH_dx$1'/v^ZZ4`9);q`F"d1v>ժbLGd~MP%m x52LMF9 EoEDYpԽ.D\dNyj荊EEg]bÔF˩ք%EGƶ*NX)Hc(<|q@Oޯr^3>Uf1w;mCja:-1_k٘%VbZ˙#G6 `q+MPU~l!.?I_Pĝ"] rT [eTr؟˰ ]\ h! v˱>5S1px fnk}sR|6d6D v-':<' pb>Gȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~)'qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޘEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lr7ߵ AMҪ1EzyNAtRuxyn\]q_ߍ&?\GSߋ2ꮺ5w"7U0)lۨB0ח*zW߬V}Z۫ܨJ<]B=\>V7¯8nq~q r㖟u-| +[~,9nY_ws]ȶM_u)M_xV hx h[K2kـ`b duhq[..cS'5Y@˒ӓdcY'HAKq^$8`b $1r Qz?ۧ1ZM/G+qYcYl YhD$kt_TId E$dS:֢̆ ?GЅ'JƖ'ZXO݇'kJՂU086\h%1GK(Yn% ']Q; Gd:!gI-XEmkF}:~0}4t3Qf5xd\hEB-} |q*ȃThLj'sQ %؇Gk`F;Sl\h)5؈x2Ld="KԦ:EVewN ًS9d#$O| y|$>^WGkDi5؁5ݢ6TǪ`mz倕nwRqR9~ i±za+HFNi>. EWz:V^&YEs5Ȭ N *7g@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ'`uy-u.M>9VsWٔA 2J1SGpw>ٕQѱ vb;pV ^WO+į1tq61W vzZ U'=҅}rZ:T#\_:);KX!LHuQ (6c94Ce|u$4a?"1] `Wa+m𢛲`Rs _I@U8j+{&g߷2KoY,`Wf1_ܑMYٚ`ySc4ΔV`nI+ƳC6;җ2ct"*5S}t)eNqǪP@o`co ˎ<عLۀG\ 7۶+q|YRiĹ zm/bcK3;=,7}RqT vvFI O0]&5uKMf#pDTk6yi*cem:y0W|1u CWL;oG^\ X5.aRߦ[_Vs? Ž^A12JQ̛XL:OEUپOY>WK-uP0\8"M: /P4Qz~j3 .-8NJ|!N9/|a|>lX9T ҇t~T1=UF"t; 8-1I|2L+)WȱL˿ˍ-038D*0-)ZyT13`tTnm|Yhi+lQ&Z!֨řoҒ"HKX 6„=z{Ҍ5+P1;ڇ6UNE@Uo/>8.fgW]kY0Cgcu6/!_Ɩ} ' Ў3)X<seWfSv!ؒRKfs%(1Lhrٵ L.] s?H,HCԢ[b C-lLG+@_$c%* _jR|\:dc5u= A@kUc\ǔz;M>dUN/aFRĦ@x؂ǀ$6%}N^ \mQ!%8j0dUo=rh>*YȴU3Q,̸*Eŏ%59sTzɟڮ2kg ۱wEUD3uKrr&"B:p`\E)j<).R&#ÃecE,dp"nPS ;44 Q8ZƈKnnJei+^z '3JDbSK;*uБ:hF ѹ @˿ޗ~7g9| hLXULi7.1-Qk%ƩJ4^=ple;u.6vQe UZAl *^Vif]>HUd6ƕ̽=T/se+ϙK$S`hnOcE(Tcr!:8UL | 8 !t Q7jk=nn7J0ܽ0{GGL'_So^ʮL_'s%eU+U+ȳlX6}i@djӃfb -u-w~ r}plK;ֽ=nlmuo[`wdй d:[mS%uTڪ?>={2])|Ը>U{s]^l`+ ja^9c5~nZjA|ЩJs Va[~ۗ#rri# zLdMl?6Uc3MکEen壹n\殸,˛*{Q=^]{Q_Trin`jSب-ȥ{Q=\;ju[}.˻Z]=F<^[Usywao!!B h/|}&amz$Wjb[ n."<׃#xuA0YGNc0%ڋAPΏe/aFc% ^tSJOk*4N"n6j1|y>^Pm:>H+Ȧ6' ,}U=̀*Eg.6_~OJ/8V ?ç&+|t><,BLqL򾷬dS{X6"Xx#-^䀕#{zК4i̎QIc(<ǩJi lc*wmreqҾ'>8E,ɥcʋlן%6U̴h]즟t M@P& (@˸lɎdYc)J]He:ޙ[j!cG40V(eW#ˢ++mlƱ6j(LBBY$hZaԢ7֭\q{^~ϣJi6idS`oĩL!ڣ`F,@N1V bp̅[KHD dt9vSX/ϣ E<K6>g?Jg:iQGiTBW8 xOo)}sc@<:/P%eX=[RvfE&=04G_"1?3u[w&~>Lk?&GOAp0K)"~!E"M)T,*_gEgB5M=GKt䜭@k:'2Ǥpfg|c]estƇW|$ \ڛօZMqXSX?b!KhN[#@ #'VE]udN.(D#ԙTY^R7 xF !'Z0eXfsYg{8,X^$goO l_1nDZ$ d}Wrʬc4MR'e̯*eM}o)m;⓸xOeugɯ'[L&mLDՇ<ҎzBS݁ y9L\don."۸*' 9L\ #]UBapՎH`+$%u]F1>\jpRa9 D$SQt\L h$}ዋ7I}C 'nq:A>plyg &_|ثJ$.\9LE jAp'nRe٦R(y7DOGJ[`*r햇vA\G1Faa\֥YmPXugn ?" \Ecy6'SD $bD` pAdItƬ"ɌlYxb2:Y 'Vm8ɬE4[ G(ke!eH!a:UW 0˶Tn* qc3n ~B /yoy )PKKHzT$'l ~Eߴj@.HPG뵪nn^z! )[EO 0Hu}Km  JgGYm`7+PqLBgɋviyhK3ݍƻۣKF Xؗw\v!_ߌlyL/O~}=48#  $ɯe5Y%"I: ZaYݦySix hoz"st-힣-Ӹ>>b:M׎5Hn'QW2ӮH$M'6##+nj&yRpG'ɵmw\I<|[g0$ZUDUTͦ1tLa&݋}|4+-i.SnX46B"gDzNUѻó2ȭ*^c;l"jk)+zaqNhРn}(hysWC2puՔ,]y;a '%koI4]vD-VRju)*eV\ǰv};2n c&!sU%oZ`pWx*l=lzi-gy)Z ןE< @$⤯-An1jDY psFfS4wml L[M%,0ц" 许"._Mt-%&9 PV[8 ᶫxJV-$@O-mR 0P=+ֆ97dc$VjqUҬ3U >h=Kvтu ^x Q-D?E<_h"O B#u;'NjJ8]r4CNZ˓[v{(A2RxRo p]{f,nsHlUENMUkjVMɈ樒->bsJtUxi ;huD]ՎBq}UNn9\mU@sm5H<򘭆ZlvEuv^2`H~*F=bu O1 hkyS.33َCwa *-ol+w_}g7$Z&*ZWRJkj,B@Ut,Tgq$ZkʄŘ͞"q[B6`TȘڒxjD{ Զՠp)F$Yc$`W!v%(B9D^~ff \oGą \.F7j?fCP _2OLi{m{W/^y"zwZd_=T$qr ]Ԫ|V2'INŌCR*ܦE-7CKF31tV!Gy PZay08v>>m M\ЖTn+°&D,GH&Biّ{yF$f.4!ם iewMP, P.f%50,Sz෸Je^ˠ^نa+ a 36( n,x6 *]ࢦ(mk*XȰICt\w㺊~R!hqڅZ2 f;u%[ ]hD5XSTumd.+>~6T<VKr-2ͱ@|fʅo.+}۸{K:?sk Mw'7~s0zhkȆH[|{0nO5?_eц0fL{Yc_<9wY﹠H{@]f G/R pþ8>ȑ#S.ECa͊\wv19&mI+(VpIYr$qc&)ER6'%%?.5u$Eʲ# tL 81NβAwx^c1(3*Aq͝ 0a0p$1Pˬ]}%e6l]Ha%}"G92(pܙ;1I ű{q>:@1zwϲiT9H;0t{.qi/N(c:COu% @c{ŜdCUG%\s#uC+]TS>*Y9}@ygM4 k09*:~t;TJT@&BqSvB-.ͩBX{14 +\MUQ0aMS}v=ϭj |k# 6 2J}EZ wyCyEpeH\-,-ETw~V UǸCׄqS)`V>y]8GBI'Cu^ \DEc<5%d|& 'E7$p ͝ޓt)9y<銃_~sB]?"K?lmRx5EMnРCV:۽ RX_e*4?\uq/" UaLt0 7}v$9 rLg N-4B{( ޣע!*ާ` xv@-'P|߯DEP|tLWaYw $ΩZəX@QF Yo̍8~йBې `}g1LLj qPڏXBq{A$ tγaS>@1~Df#PG( nxL7ޤ2R`]wcB,nX׀=aSWq7@aWB\>wk9>+wQYH Vh@7@ ˱BLu?!Q:f-@wK`^"7vi `qm3x_WB,_t$=ntg@`?N )LEd7'UOB6 Rwx8ՙu{'?\\Vyt걢:=bP]JnƒEuGD)Q7F/ri>Mݒ({K0^Dh7]I)UvIT!2ڱs2]]NLu2fiA¾yoүMLaoB]p2l`y|.5^-P˾ֵQðNBsK&!QCvu !Np~elP!PC}BTv1;f*Tw鐚nB.XWs^mzghRpHԶzrN @Qvt OmS%ޓgk' k::UU9RѨ%G^YykBw1S9 i-xCwA }n3uc9o]F<2?tB7BAKt!Xv<4X@(etл= t6`&-b?t\^hd^_GF&m7[::n( ]HH 2O|}9 \+iOei J_nX >m0[^)@!̰їd~IΈi/q@>~h RZm~Uכδ UTmLcqa U ,p+ K'8 xXuL`,BYBFFj(gǥU Y2! '>I;rO.+ߌT"ehOz?WvkKgdiFw:x۴yòٓ:{0Qxś*1I b $I8~S7,,#%<+ `ùڳMŨّ_疶.O?% #R6bN`^_tVQ0qõt(^L4\izΎ-%ӣ+tLfƸPЛމOMКH(s _Rw)UjA+_ =x(,TA괜b>az~HQ<<:w};vC6=pq`GrQݎnC,Mc@ s`\'[ֻ;nXc:*,Yáp e7>+脕fKt4_"y_gh ǫ.YOƠPdϲIAưD9'h&WY XYQŘbyTYr paU9a 孕Bg5E`LQH?ULE*t6hݘ` %boYnRݥr@f{Ї6%.( _G|H>.ff ~zPD1-sdbH`q8Xxza[&er}C/| gdz_=dG5^ϲCX קRekUN1_]ͮy7QV<0 W1l @oLŃ  3s0Msa9aa(B(,3}'F40 2&w lhP]㓿=k؎sq'Bӷ_̃8u1'~Q^>(5`ѬQ\S{W266r jsiZOxV(cYaoQU. %n0LuQ< ҤkKFN6k'/~&.+:03!FSs6?$O3hs/qKsbR>t!s3@' srK>p~qEUŝ]-TD(r))ey𡂔bOCFE?'~By)rw4N#~TMPEB+%#NuUMXȢ@)LRTA3 s`CbKe9ID}P( lIA:{_5|F+B5a1h@ 7zhyf"% |Y:9ׄǚd.UQ:|0[DxF9(O~_L˪H,CYRVKm^B_|K&iD%IG0ei(]%t:|!Pe7,i?+a1ܛ GL}*wО8YܤS6k ˷( U_?] 6=p1Q5\K!{JV7|/Tz))A@wNT~21VN30{U0Q=i[QBzJ,uqXbLݵޟdG W%dzҲU,7IJ zP$dR%0Lg&]C7BI!Ѭ)\fkrڮdf1oSH5lj%v#wI '4G1eIiOhSHrw[Ks \$( sODPjU귛[I6P̴j"5CቶVA6T 0`(`hdH$6 aQ5FRr!V bBCZ+|J& r5_oXjs$G\ VT 3J B/2C*.\t&EGYHv RM l#+,ق JrTQ"ǫ-Gr=x6 :4O skxC rR˴\$IUvG$~&%ö_X~J~\i#Id [:ĵG>Ds <`f0|o!)I{=/3ӷ'n=6iZi^}_W| >OSZ?vcv]:WV^a N(\ݼV.ڡW 0,r'~9UE v~iE *8eO*aܡ[YʟEM;cwi~?o(s*Ca!3ȣ,UCX])#I@XJ;T2)4ηSSDA<X o 'L;?1M wة'a䙟A:l,O1j!af>8Lw;Gjø^uWcQjl\kJ{kDGL+8}NN6ʇ ENJ?ׂ2>9TAߌa zM*oN"K#ջJB#Ί= XƋ[vCy-G_DVĕϦ+!l1ŘUUfb8fe;;TaSG 10Xb`?.QTK' {jY%.'Md&/$䑲 J5.ZI+{c+ếi${pPҕ1gn6V ˆˌ=Ma Et|f}56Z߿-@Cd}X#Di`#s뵵pFD ,ȗcd1F ;BQ\h&2<~ꅯsY ܥʠ4IWʝ|nȎ^7wQ4D~ 5%Gr@T;C7xA,[''/sqda׾zWyN(7%(~㍻ͭ㭡ok+_~6Ke1v&BQ.b0QiӸ1ZUYF Y) !d1>wmIT~?c SRD )K$2Bwg-rޣzuթwwsQOAiic߅/< NWü{?}$oiSeEsc{| ,wߺA18L޵kFZ7_J?+`(ʏys/x!>ghYw??Ij ח?ῆ5~r!]7ϗOԐ/>I@|㟗w{.?iėi0=ayH려aWQ~/A {4ɗ_]qY_O'=i6<}ͮm˿Z7ȜdvUa΁\⇫wCb]~OvCu_Oz9^eZW3y8壳uSzǷ)bnC a׭5ML:趿9 =fx%dUVHS{쿧H>u!Tҁե ³PF OS%xtCEy6 |~Ox2[BgwqF< U ;eAZArS2&/ "=.X>3*-`]oa OTΔ2Lw)Ax O=ϱCqT#, /$}`ۀ$GY76nJ*5(D+\1i" Lª v~̢L+P8𘇃 F-gӔ" MJ 8yj=`l٭YK%o/o!ʚ Ջ "֫)ll%$.ڀ*0MPB :Z(-~ /K{$fCj*.8Φ@!tf3!*q_;&& j,vJZl}&OA#Y=)BJ`ފ1DVATʈ A& )yC+tru.8霮E[_G 2JBLEC0K/BaD p.8zVU=!ʬF[.r'hMmCj >YO!J7r"^B4K5$#L'YMPH52.>jKT25WiZ5_ 1EH8ctюV;ч*fEeLs \b iWbibK8"bTc `㬃*E1 vi}Vll,5&8B+1ʂ,J|eκd-ȨAYJBL~ҜOO>'h^\7l܉l G h޾gJxs?E! `&!Eap_~6b3àa;VPx:-H: ?.8N~[jlRLI Q| cr`RjhƦv̔w35լd%!C%^/hujXR'ĄcJ_uE&<Ѥ%,rK$HǻEbgBE C;]:Li_`gHRUc Yڄ.ҞIyɩnFo;)zɾ{{aۧ.8JKY}+Z?D6M~L}mZ< V6NX, _N3iO ZFoA0l+nj¡GZx2֝"8h(f8GZ ` > &|1L(vs%ΪԴzFyEvVp]=i n:,+WA׌Z#yأ]K6:CH:?}lWov7S oOlqPZ0oUkG-#IMn[_xx 7Ws,e"7W-~-@Qp9u'X_NAdXtKwI!Oqk%n{$eCe ҮTd7ҖDsu.7rp}?-uNUf P͢3@-ê\Su =Sg?+PSg ׻Yu JfCXڥ)fh!xZ!TĔ 8fÎЖLGX"&mo{?8ݼxnH‹n0o9z7Vl1\ K4༖`K24R>1ل 6 A>xI$'t#,*r"4W)I"AR΍&E` \Hq u }9 ,qޝ ( L?,.8*yV#?mܔ% Bダ)Gaދu10=#Щm yN=6[4A=vtkL}trʩr#K9?e32fQ97s`l5O1BEB;k<*hb!A1ϋ.8&>uGjow:Ǹߣ 1h{`7T^YB:x{8sԎ_&5y8!AxdkkKx1a5P"|gDRxU*:! %eB *1E߯S~c:8{GDoeqzaMp=hdOZJ.mRXiKCΪkFI::R6+0y% /ĦzS~<9z`J.!8f6k:;ǽzeoShiPSYnrYwOZS쨟ncr=D1 z_f^1R[ֵ4i%~6Ol (|zԫS QrU;-rrv2ʔ"inkr~E<NtH?r&ʦG1xqyC8ܛ#-kP5֏ j$.Z `x14 sIU1[ҩnBawhUx6/.0%$Z@# cڵ-`w/4tҟ E!Yjt^U- y0.cMM-&Is]jވ淮~ xY{ Iz1{V<3h Uy^#n/HZ`xd.8Ӽ!i:5 ->0Pj Ggg$NnuTpݯiDȝN ӬnMa:" QSG.q[. (8;_D8zsY.pm[Ji ;j^ĹYrB>t"&8zw:Ok<аLW-ޝ.*O: `[umҸ< ]VvoF u6f)9dF;zSg>_uϨ&OȲrth$O:"cػ6nlUݵl-$ 5>hsjn~9cHHvQ؎fDy2Ub*lMa>lkc;cq;Yi]ė`koa̷b=9fS7 q;px1>όHnn$ ؈ 嘰4vJK݁v厝lRe0 (bB̭xg*-m]_=+p+U^'2|[;~ &[}6چ0|me]pݪQk͋gϠqkӣ@W9 nZ ěVqOv! LrܪLjIq{:r}xh'pA۵rUYVLQvjԎ7k@9FDU.u\ԅg`[Q֭I3B8JڼΡCˍ["UZ =ΝM&ǝ"&ᴀY[cϝhHirn>kN. 9^I^8K=gA5`DtG\rQ1GGo^/P/ (S/iM5}Ss>ёWCȞ=ʣ(suQ ~1(o&& ǨK2'Ԏ!?T! @aгrrC`KGb@b~^%/ $k@+h?=ەmL3P3hU`]xe&oWi5= a?4(\{a:{<+ڑO'>XLƞejB&EH-#Szs%M29 wW;,}47 [G/r6qdzфZkdz9KL_.]&C~qY:≛EǨlֽ@ G(Bwz~ :Ct7^XѤ;"[$K3z_7c9%11ACnEf(Gh 6!9յl2s5|2 .l:}haڣ∅W/|OWӴI8R QB.ę,Ths<*U4a@Tl.0*FW Wn&y=^;xyԹo81 ʓ~r)uT{.FGPikd2 Ç=v.8{)ѽH'rV\^^؈8"q1؛Ae,&.;(?xI&ֳw!'MSwX[}Uⷊu,vԦD૗4PZGY衺)nh <ݤL'HWy|,}Ё^* (A{EЀc;6UWwۣ8{6Ma6낁`РHѢǾt D: XޚΓKe[:OXNWϦ~ef/&wVFJ&FZU#dTg'.J%UuRP?R3r0!d$hp~_)yE']`9~U"v/7ժ[ۢ/|,~˝ൣ/=3.bۥ~'K/Z<*o#*4&{J0-@0\ۢrAhutlZCfURa ֚_I >aUnhkB'fT|nPBi %$nP+O@uآXѳi4!{ ۠%4kOM{%_+AI&iPF@30&ie< WIh֥Ah2Aۀ:U/] ̸n|%dė~@n.=VaJ-vZ5kU B*IiLGeAq&iDG(At.6_A3D&|ТIrКa%e[ aUkK֡_ݤ/_I .w|˰,l{rōG4@%46d g,76& ]ke-gaB&W Ε Xg2# grB;2eʗR!Gz+Gvz[ d!:4ϭZJapʬVJdL#bòVneÆ8:l W6;,Rq⸠Ru/]:s$!sQWY{&,%RaeS ei!(R/K,qHq:h,l+ {> 2ØკtEu?RtRo|UCsöɽ5Tbjj]M`e4ʉeW8 }[m+@ +ņV=1JOh>) k.Y Ț?s`/hS(;۴ kўMOn{^CнdOimƖ%;YTfX+w@hKy9@{mC`!7;98C/a"qZESي,x[S<3 _Qbo}cZoV / pK%nDWae~=~"ߺ/=%dE{6ÂpF6-OnȚ-O8RIp?+wBJδ0OUJ}mЧ;i9ٴ~ G rˣLG"O[VUŃH/rIT/R.˨0\K!4GQ%\pG )M#̣QڹPT` _tX Mguw_|e~O0n2Lj 9cp!qpmff4Mi3 s9'w%2|ܫ.Ob_ٴp!5 Du_gu9UaSC5dhܷd,KiƒR F;øD ;dҌH\b0jrAKc@T1d4d4*y.J~/7If#IN,r2% ؗY#8wĤ[Î鷆]9/3>%[0U0ݷd/ cΑђQTpD&%D9Q) 1TPTI]0ZUvHv~wɐ\F߻^ƒcyj BSsg.Q"7\R´b2+KFӎIFa{tY{]]lVWrt#tV<[էJhn+/8x p`̉, fP!w1;z #ȟٯoCoO~|wz>}:0zK7:omU O^'M=P1Z;6沸m4F:\͡Nue lk7o,hIy@Vi& S7 W'ԝ_S_|@[Uirn|m m߻aݳQ1``!.W4|&Ť[,S~/#ͽX7Ӛގ*qfy`B(YkkzUQQek*,HC6 ofCor-djvܾ 5`HݍPcﶍ~ɶ_[` keY}>/\?{*˱?뫓7h3QW\IsQzK5x{ g ֔XyW!o8jsI*vO& s@w%=-MC${pZSyib=rߥeXnqibu]h|ј q_[]ԝFc Eլ%o -wfRq_}<_fř_Hhpz *4 [7|#^I3Uu:R]1zo،Q0+?T\e}b*/םͩIrIV QK0:CM{% {ۍ'L=y 3߻W/߲W/^Z7ۖ;6˕"Bͬ<5SL$ukA\u*r\lH X_cP?72`՟J8OJۣ4\4QKE՟4TLn,"0sJksE E엌Peu_W8!-|l r5@6i>nri{4jL^f1̑L*wa Q J,8BD#2M*\2eNxFR pnh* Tb<( U3mЙc\Voe`X MNYfhP"2m0s)iV ЌL! qՖ咺& Ym7fBG#@D-0LVI)ܤ\Z8`Z0;T@!Bgl e|`W-t v wH&}' -pG߭RB Z0mjE7:9O-cRD7>j(㤘o梪KL eʾscww\nfQY9C 7O̝'OVrV$Iݵ%! ß%pMP-ʱi;x#7߅޸(J^qU]Ahszb*))Vꩤ*)FI1[QXRnRt(Wyf2mkmH +Q]ot@bMrF`cf'ב~C=9!y'IjuUuUuwWK]Yǭ1"sYkWy2^12a_˜k!˘ΥL]Z:gYƜ0] қR;iMKَOjfe1n|sM2[QTdr J=z1o+>mIB]@^HIVJ_jB/2ȏ ”(yy[VXib}GRXx_,Y=A&%M#}#;RQw9wZT0 C= 0gwBNO+ i<- sZ@3y@y(r8#\JhR i@ n! Iqw\RuhLЃ1 JQdhZЯi->LJA @%\jCDX 򥚉'#H-zhbrK![%{YČ +$JYx Yh)HdYq %ܐ8"<cBH.G--y;A ;6IX]ʜ59MmQv.Y+ր ̟ j!%XMh\1Z%nrC4<OrC4#2!Us&=q E2ԵZqͅ qc9" ۼmˆT9kNr]@U)_$3C8r&],ZĒ r$w黧Qⓐ.:}%l`RW[j:A)ʒU%#).eɹh.1wy2YYw
Kօ9kTf]10-خPY6"0.t>9!΋Q,ttɉdBsRdv %; - ].Ȯ.B+Pu Lt|K݆+[pT DHE/i|8 pxyg>͈p|c:0&?ؙi4i8l`Q5.;G\yar\f3/|3[*TB.J2 /Ȼ7DYiQ9#|a2 JvgkQp%A[CAn82甧M&j8J+{d%ַ@ LɅ$R|'-郦h)*i#* 2eɔ gP2n3mJzD^YZ9Q0XZN̰ .U[J^A>, pN'A!Yr.$Q*eVH_vdp[d'+}nKs`s ʎ|њ..$Y"ơ!<ըh~6^!J=J*^ Q3" &NDϙo`38F-;e :e yn~OQ.[6zhBwhD3FKh&}55QfB-Mj7&㊁)_ ㊃UZe#EIQ#'$ UF LX82}\ͺ2v@Y%˒+FRǛum@,@C3+ֶ33+M|1X3F &XVE10HS3F5+ V ]mS[)D]l3JL{ -rwQ↢XXQ0~#q/ ebg[h4RԺUB:4;;uƤ`zyrM>Wsq:'f5x0wa69:yvz?yYרaSkN2ퟟk,z^,u7Y>s]W*H':8݈ ,@J9v.,sO` _8rDŽu,w$g/zY7@dWaO+9?+3Dig61.eTT2_k99X㋙Q^oՐ$ZjW:w_ox3TgG?%ٌ$fcT~$'GKh+W=gEꎜDńaim6W~3n 6 ?,;u .zXyqc'o5oP=0Oxp֮tZij;Aյ͈vFMϗ?}/CjԎ|6۰J-iCƈJj*(]"AIRS69(Nx2;u|qߟ3k Zڕ6vK'oص&C)@2kM"uSvNDz`*-EN}c.QhrM Ud];9Sm>kl-Vo,~C騈Ξ~6 ٩k-G}  0%6cוA!ߺBʟ< 7\0ůſdFPK/fY.҄Zո*>;s̶&.[`z)S0('g'4fp~*!o͟<D,֗Ygϟ?nrq|; \5)!R}~4_"x4v?}ʎ}r2 !=}>^w2 HJZi{vb8>޻N, 'ZzZa>4l4 8:>&\P273oqVe:U[Sb7|:3W_{P77;6sI 7ԙǼoă*ʝb_CwDMOZVpܐ#l$< OstcIxWuhr.//e2e_d/a`l!E[d  fٯЏo"N{u駳O;v<֏T/<:u{ӕdՏc¶˫﫿ir}xf.>һٔ^Iw{?6O/,v/TOWB9eAQ!J_3^T'kE\\j4 NFJp$C8#Oio&s>VgO+QLZ%H4h{и!N;->Ұo߾b*}D0~"#8T {C޹O@\pp@zj*өI~?~!dϔ۝ӳ'g\mµpJ:8>nHϚT1+)Zƫԧ R4$KLYH, fLϢs:UƠX۾HlxVVL)V}`%ulT)pq9&nw 7q]k>^ /؝^)MXLWꝝ^X0Owie-C|f~-ؐǿQFoxU늸DYL LkeHBG+Z8NyWrڱUa`Vj~5}] 6bU \?ލ,OBnKhR r^m4w8 ZookݩsAה0YUQ:RP}瘝Z]Y .lzn>xC}+*%^>Íc Dͬ`r<͌|g$ n*IVHfv1/_@w0Pf?e$e<.hسE87*n i`Bq`?NˀVږk݁X4 5@)-/ĥmHʹyV6=\ֈ.H]肄bM1ͭZn|'uK  FQk ɵim uppM}" xo7^lH掁̝fêjPV^[hGM;/hG*I6vOF k$) of¾lfQ7tz$`K}{6Zk;Z}{]UlS?ِN.}O'7<r=Cb І V_ڊTmui!\'E*p&/I\l"8mV7iˮ~i)CusS! g>TH}8 ސ}j}v}7d_PYUOmUq =^]媿*q: xU6NRZ(llb7ͬ.`Z`)AsiWpue ZNpJ[[ܦ=]QqO Gm19?Jg & Bڲ\ge&$7^.pw1dZ|]syR4F_xJu[ HdI_\ݓX@ %Bpq72c69[Yh t+2?;JkZ~77FpHsm$WWq9vHV.`%X6A7gv؞L|/fjɢ9X%6*mDНIk'; =zsq k죾V9m߬zh.#2QTubSi= 'H ,!R;! ր-֜IW|u)$vnӉ[M 5nG[9MԃAqʩ&6ϝU k!?L{Y/9IQڵr9D0tx'"<'邛 tI̖,i^jV@ qjcP{}L̖Ch;u?6a[;9=5lĆ6SuOuݗnԿ߽Z p=kSk|Xx. ^K,ꚀeD\VmQׄG#Yw@~ /_UlݏgW3N"4L%xx{5//bL|/G-E(6ui-"ѧUԡ2܎TnLkw&cUݡzwhwٱ6\ҭ<^onjG]'+Ki:>֭uZbbV\n]TjU*ۚ1NtYη"ު~ ZyyԛV52Os7wOkLuA\ZS%d.3 *|!ppBrܿkOFB[V9Y ms^>q[7WAW{WuE7o\R)?RpsĽ$tN |0{Mw38\>m7iMX?C ʉop,Ӂ*-vzAh=Ug1Xl !Zݛ~{#ݛ>q7icG|ܻcvh-++ Q)Ǣ_ ANUQ&aYeKeD.V0i{,0Y,gTW y{T]2dK7b6T}U}LQXBP*BA-gg-_aXy?K^f|[MlA,è$d#I N<}8IҢ[$_ >5QmG#GRxrTYy!_fAK㈑l`AaE ֘˗zˎ?⫦k6>XK]E/7#ϡ  .}byԘ`$5UB5^6i^RTf5Faۈ&QQJ83TcwR-Qp@l>cs725v!MEi<2#-nk΢Ui$X7 VQT-nٸQp_bi>2#\EQDΣH*ҋ[3s{kTEQxJw)^ozJseQ2EYT+dP;I~2Oz{f^oxmT>I՟Tx(noq{U$?bL *Hq|q(nuOTY[޳t{E42%L٣wWMШ 4VDJh=S%T -noq{ 2e OLd[sq{H3Q6,i)=_׆.,RVQ(_/?U7)濿1_lfM׎kRjmm,|V̭۶ґvbyosoʡH_`@,uU X,Sr謴#U0z|\ycʙq#܁q WCӾΦ,j7KOŊ2O^RY:' r SeklTڳfD5-U5IiIpFCk5V`1O {/&wlF-eU׫|$M =; ֞dzl\*4ڱX|'Bk&߰cJBD ȇA, ZcP( ڃA?.V9b]Y oKРZs0IF򑾈ٝjc\盒ɅޘG i[ pbF6|g=ؤGE]IzG3ŇЄu+mRMa)W:7~%A, QIlo]xڋ 1B@b-<w7_kcԺ1 m+JhRZڦfv]x{F+޵⫮uȥ֑?$}Ol0rDA#lZ͋֝V WhۓqWR;`$8{ʒ8 KexuTu"{`M@X`$x[X2:l’gg/.Aed[.;Vb/T:C-{"]k /jZrU?zEUY$ʍɇrX(*hBcQnJu{sE(A1/ F 8US:_*akT/{w =4%v)p F ;*m_v—7kn3%1iO'xI5 а봳[V MsFcZ /Be/ƦDo F )١x7uͱyw,ި !p5 w›q#Fl}ܹH+-  :pe`L/`]硒dC.%3bIۦkʖ%d_F }?= !x% 2cF Y!,-dZj`Gi!/i[[B\4f#<rt*$=aAQ#&iWr¢q#(_iW; U|E+1nZQ:HC|a5HM*lͮl͸ "s PB]J82n# +I5Ť]-󐃙B}ًUI5 ;9+%z=lUu zH,a7zzе0j$89_ipxiyP_6U%i k-Vղv;"oBGσX|^RD$XVj M%ՑFm7zMd:FM#q4T; iIt̓Hbh -TLx}9FJ,|l[| q)Us'jZ, X"`_$ŠҖ*O|@UG "L,_j.1u.h16W.BR:UPP |,d$w&!!.̈HmR2n$L*00*ZE7-L[\Ph[6 I@`$hXF'u@w`hL-ޘ1KOl$@#H>ЈShmB@ $3/ Ƒ#MJ&Z%*_QSâdqř3WW`ozgY¤E Fs @x_q@L]qp{p6=H> k| *Ak\ Y~F C+ l![FVҲuoGP5k&/5\b@x*4 %:J0 Dq"Y: DcAD8GDO$H>!kK%::vRU5jjs X=-R# $3eg!n!4(9BK!d8 , Ohq,lLjz5q2?%q; @l2`w I5:jAꨠP!mqy#d2x467G,Cabrt L` 4nvD|@8MP,»+;k~}moo+(ewOrQ#*PM}rE*M_4>݉{9;4}zU8A@`$P[@j)8(/3^!68y")7 (Xm;Ӗ%!jILQ̔ HZ S !dGx-A05.A6V?o8`ؽTQ4yۯ_E?./kzL[:)ǿ7mwqrȯ~ ? >Q.>A~R}OϚ{x'[=Mr}K|iJ"z|hr_S<#}(dg{(rxy.QJ%$`{A,f:8 &K1ȥv|g_I҉:?ccm'}xv7r UowVa6+^}KOcN1jn44VT 0yqxl:<"86Tk]&1q0]L⅌A??|LScC`& \Cg0rLӖo֍][O}fh7&z)IdI:SPK+7˛>զ}_;5vdkuܠoj~AR]_~/[㗕?ӷx-]"R$3WRG2D)u揖zb{ShM',颠b,>hl8I/XጅBORqӥ@ec#*0`PuBP)?*hJLjiQD2%N)٨U82&\ߏ*+u)DjU٨ĚS%Bl[wxkIg)rqr˄Ε[9YfTQB1EKp,̌291{v5S x2p`[;  mZ M,TĸuoMwtg~1L&@J`?G _麅x4ӵl~b-ꏽ8bDSEZw` $lҒafL+?w]ml#j/y fKޕѻ=c?b0t+*L_W0kTŶEu1K]b1*GƚZl"?C mR!.PT(SPhJ 8z P`qL" P uP@L.bRA 0Kw˯ެ 9 kl-[ڿsQ,q>c[ƃFeC{Кw }#ClFFb.u= ~FdMn {z'MnNf>7翫wxS !$c'%XlK^K\pDŽs]Gk͕=Ѻl@Z T$fp%ʁV&r4ԲvkvaHjx|+̅Sl'0b6!Gk+k)_=aa~%:eSFk)'ܳ 21v)j2G@Uz%arb$}SV] qy c { ¯?`Ki]6r>K%V=/SKi H1}G*Aq3~l^%G-ֺҝw(^ʨ>B4+ĘIJF" WN7­}z$*K9H] ^nJr.`-b#F8&goMVӞ9+2B I% EM,Ti%J!#kXqcc_y_.iRNnj)`@ $l T~6^0 THt7[ zx׶hER 3&TOC)@9_ I0 ;XD4,'[ >{~_|(xdI˄)Ӕ!wq'SwhHZ/֋"q/2 J,7ŲIܟ^"b.dHZcrHzXRWD2 ,wc5H~l~l˳C7BӔۍ](5eQ0?vzɕbVV ~ ߍI ݘO#0#幖g8&K=ٮVzՌ훱 Zm;AK;(0˭ |GwI ,o\m<4;^_O{Ӑi<7o kM%֙y9sOOxga?˜Ѓ޷ .d´Db%>N3wAԜF.'owV F/ C!hшbq5h"پuEJz(oǒњ3 .֤S19KrLT(cSls/_sFOȿ1Џo6O1[_%5v.~]3\L*(;Fd6Uf Jn.1fi:SC+xX%$)I^ ]B(#WSQB)g}cA9sZ%&9&fR*RXiۇu=+xy_ +F`G:~|4gi>љiIq"{q|/n{5)n*q'%9F9FKr:b%5]*cu8w]8w7oI.lE2 wP2%p.lujS8"1 =pnqnQI%9dA.c$*8k{c)'bCtm/t {VU?[nR5|ôww*рYߟpvyr38^?.q|sL[/-=6̕{Ӌ?|Q9O>m//yvl=ߴ֖taJ3PƠcq;?0,F.VUIuBF9KrL9KFG1ɸdtcB#_`QڌD0*)0Q`T#MF.jºXzOvz2OcSMCcEb GNJV[#c*' džjwP*&1q;P͘Uq)ZQ5&D'jL`,b949`aZB1W1J0*d7Kjr\^,P10I\ +4ZOf؂/[~U\pæbcgmpm{W"Mx.&ޯbesݷ_sp}]5 @!cJ]2SG8Uk6. МiLCNKv&JV@G  iSk8;+yMm4m3 ,8 ox)'ƻœJTc-T!S.T"Xrv`؇B- mj;#Dy-R ,PS~_` RY`*H:4%kд0vc\w|Jm>fh{ߘ$iÓ!|'IP:Y=SN|,WkPݸ[" 7#[ vEr4\\T : Gٺa^|t>9~ΏMI唓QM[> b0;5aS .uR25Sw4,f.;amEc5Cq<9m ٥bݞ\*dR)]jk}` údUu=t<ռ0]uʋ"g#B{_V)ѸF̚!W0 czpSY`~j vXSYZ)qʒۤ2@<Sm)AHB⥷8R5m eWn!W,1jL^R5E 5BM%E0KrYike4z]! K;&KqYIBB)!^:G.eYwu= gۗAB]rK#M}fc3hj08]Vޏ35ΠNݘy}̮g?@Qt(+RHf~(v3z N׈\ ]r~Ud`g5g8tTXKFP.\<~5N(#M`L2f~<ӗIRrV.OlbTLB`bT`I<1F⇝U3< Amc쌋>>U=~}R2^Г͡k_dE0rɸ;GyL ߸`<}cУﷃ{tF)q?wmz % 1.<.30Qgě%*fWX N^j`Tz*enY=+\j{Дt A֛<(NF?i0VXyu?.Kc s$XR 3 }[KjYjR)sX<Ől 3*e%d=FǤj$%J) ζ^ቈf꛹GnYf>*!d-҆NC6#4)hR!TP2ƊBe1Vv59{& t\j |jk$bOz۸x yP֗g!,qHR%q5Y鐬}DGdI#dߎs\((FUNXɪR(#DA)IAfڳ"U/cPB&vju})aTnx);A)pGYGW(mF萚DC!-rL VؘPR@"Q+59KBD9jmE爐1;EQOJ=i]L:%$ ,ڐBJv')0rCȘEgE†< V&mY)dU(! Q C꒣T"8ˈoh\z ,)K.`dHDڤ xd!.v`QmBg$2ID(6)e$bvPDdp%+X(œ0(1"di/$@ DYS2`k.֜ oXfH'mY4gi&HZI(@ RTi޲({CE8]GZzBTP,$`?WӬeLC&lZihlڽgwY" iWˋ˹)`4P` ơ+F`%GB(cӿτsAEMAڳ^y(Y58p4vf cL]{uj=#}IiA,@aPB^o+(69Q͡xy%@D;=DK wPzAAJ@RGTiO$+Smx꛶51lZ6 +dO1 + EJEs\(\B0٠O:.!3U C 3؞P(dQ)2QMDY 3^EU,W (QLG/ Ω) X;uȍƙZ{i&z`HkO,UiLȤO D )iZ'k-kDvkjS=x3$ʾ,;j*C Y(m@8XB,gaC#-ΐ& =ʠs!1ޮt:xA׿`鶷]8z}zTX|A߶s-+\~\+Ū):].ڨc?K:(`F<Ǩ6hɼzPdu8F)uQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQue324Nn"NC^: vݨG4H(Mۍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨj:2QQpƨVzP:F?Q*Xq7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tΫ1<\~|rS[m8~/׾o׻?bQJɫͲx`x~\'~BN}RC_7|qv@o˰ ڶ=Wr&` 3K$p3kH% X3q{YڙP.O;c_ XE XMY&f yϥe-siYl<ܖ%uy#P z`+,$AԖ.Z\./?_9Ճw#[L~+3ireWj&`!5Vx`U%;K˲%f}[fyl5ҺUBNͦe:Kr&`ers k\Y,[z.azr'X},Ft\*e`5 ؁`i6siY":\Z֡,us3S'X:ݣfV:g*Y陀,Kua\E\Z1W2ӁO=Aw͈.H4LrmԷ_pU}wJQ9LatH9 o`&T7o;ۇ~OkMxe<`wKk3v.`gv&`{;ܞ`ݱnZа$f-P%fPNəWNM̲ 굂}۶gwawgazzoǭxe`pLfJs  uuE%)؍u;>p7Y9/ggn;wv-/oF!? 0==*u9硥z헡47)m/CfhWYĒBz;p"[a76[{f8>hw6YWCn>~for|oWr|mI|?߾4μxBFoi%()h "-LۗQ:msmeZҦRl[_;҈DwvA&!6{*\nr޹q$Y~^yvf , A^G%R[}O49mnfB?HhTVdĉ*L)kWBSã濟{MlkpsM%Δ.ǀm!ZmY!1:L} f}MGQֻOc02f2qGOXd`*?z)k rp%+G>n;r16\3wp3dq$7;Y ¡FzP-6;fO vkXC4gWb仝a@2WH _1{#b) =lf)Á3dZ\mv^v}̳͌RLL=EHx" ,>.z2GZcdVµG01/YM#:-j26%?!!ŜkU>gNLŧk~k9QD@nO񟙴ߔw?ץ__)F+ ɚ+"Kr-!Z=8~H)TF||<(dW5Jf1S_zD7_!8WdAJL@gAOmZLT25f8L*,WhRF;YSMQdb98Z4tx%Ģ:D{|>P]z0Ó5%^V-MjV#{dMXd`L -\8:)E}9\҄nv9XƥM É.b]XF8,yuԦq5F[M8.,nY[3^Xk~ı+H᛭K&$mc6>k59ȹEۆh23pPqU@/vpȝ@ohs& y Zsh3tg@ZD*yd2yy ڄf3o::(_{F MW4-~*\f473z :-(&Œ0m'۵=ܱ}TbwjGxMACېn:HX,bD&\ti?0͚Fd(vo&$SAh F޶MT&Ǔ" pҬC0{T,g Q!]`(68IHތ(Ey@?Q֑$(ĩhHأ8%׽)iH4Z>Ί ȕ ՗4P;hGbE\nXM(NO:Ƃ.&bU6ZQf^ 4Dmh2QӺG Cr>˸JLJ{m=K.5嫶f*e[dQx,De60iK+N ~T}k`*]0ۄ<&{k}hAb;~X)v@^޿ }Q'0%\Ylj+vWyŶg`dFML/L3֑3zw7vp[X97a9W@g՛iPBЌۉ~*yƯ?HbInX!;KH ZM3A*IA>zf UF<(֒=ߦ%1A@\v$Uw3P)mJʛw]G?y~v{u9~}z<(I ;d0-`]lыqmpB؄.gj@]Ue-g%3 B(Y$x4v/`(l տiSXz7lJ4𰗨pP//h7;t8h7Z*R ;X*Rm*, eg0m[c ZQW,H$>ɣOnă%Ӈ95Dym50腳/U>Vڢ!:|@%7?S^u[EXSmȮrF; oRtaV;52ڴ goLF-$ęPk蟐]-lt:3<<>H-[! Acv갪>u@>,P$tnhh-B{ԋY!W+ "#]2ib\G2r:qK9RwIon-;. bS9McRCf`jCu&!:`kk&t$9ĕé -ۺ(`z+-*\0/A{{ r cQ/9pφI%;)ǥ:\VoåEĤGGnFs˚wK/ wL/og5;`wq`Or_^ܟ?\^qp[~̿ˋ'޾[bKiV~ʱ›?ԯ q^{_Iv|;ԑIOS:"xԁK #j]|TPu1RPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAs:֡U[ ԁu&\ˀ:rn=$h:_DuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAuQPGAu/-J<7M:yPjSxT:_%帍: (: (: (: (: (: (: (: (: (: (: (: <Pɓ$\x|q1~jcۻcz%ryzKw9vsWbïww滛7Ƚ?9!  m"bRoZeesEr=bѰ_D,ɣjX#SXDp!v>羋 Q8-"V>i)"b21 oY3lju0ɞ9u sm"b]鼮>Ê )JĔ(RJGL}&bs6EIJ:o6#*`K#5~?1Qjޛp7_ܾy'w_]ϛ_Ebq%$2`|fXAYY9Vi)VYٜt"b98oG/L|ާ'^,OtGH1>yX2b5џ9f+/J7*+ {ؘ-"=MaL )60ḐӎKD9EZ·E:=b=9"byW H*@΂Ufѭ?GKH ,1"bQmbmUiu*b5"b-"6ZkL/6Ywn&xzEIJ6ܑ\/ E"bUǁoUVV"vl->AlW6E*b̮!#Ğ>v~/~sI"{Tis;*DΟuTjF2ҳNtT舳:*\Go3*}Xc hQ`6*XuT Q:*\GpuT Q:*\GpuT Q:*\GpuT Q:*\GpuT Q:*\G?fT)C6i9:Ӹu 7e@}guJu"P (: (: (: (: (: (: (: (: (: (: (: ()GBi99I3:Z~Y `,kmÌ,uѣ#< ֓rf$AVpvKjXBA`5f `-ɦAyeFYG>QFk?2a` `0 RLoQb4v)<XKAl^!Ajcܼrrr*lHeȭSΏ5Xpdw&Qr;L17 ؈Uk&#>Q9Q0ۣ8#p3hX+ ڡq)V3~ X/FY f < QF6z鸄nk_*oA>č2(`'v<X y6l{2n٪0QF6@t\,c?{Vf5q;X ,qڪQf3zQA5 Gm8.X{)Ń5tAZt`4Q1S% <Gr(`#Y~ȑq+֣x 7QmkGnkr:,Vd3QJ `c7 x#X~+#?h,15Q.gi `93Qum|v Q vSW6TQGeq1 D5.VTm4J̺HAMՑEǀ 6t㲮'PDAjxegWy}?~ypstͣ{u q7Cg=)h~ -'$&͡?sk?~/a?X|W歟{]`׿zK7#5M_Am؞I.]7ۋ~asy6:l}{UbK^eIQ}1gwopqӫwmU\[!EΫ<;[+Kוo79X.,~_y==ʞړ7X)Wp5'ǙU 6ab] |By|^0,+yӢ毇396?GJd~IBB}bKB%Yn7탭ɉٝ9eH[w݇},>V'yFcNvVMO(u5R =3prN',kWWsf?IfU5PlJJUQi)9UMU<0wbIN VUE@(.3Qʌ`ax{'.@*kQEjI̊M {`K!BMP![%kdDMԴe]^u9E^l&Wbhc\Zޝ+5;bf0)טkRSLj5вTۆXzBq ٝds1׼RZ䔔^~h`&9ѣ_VW.YWbΒROĸvẸQӻE4P&sOV +%?}4L4V9{\wEa KxtHueboBOL/YhSŕ'#Ynj(Ȓ1'CoM)R Լ*-hV-w򦦒Zg4j(%%VRHr=꜌%:־3z_ӹ.Q1&ZtrB>8: AOkcm DV}W"ZKg8|.j() HFUK ~IX(p֊1m3N\| PI{mwEl)1$Mq{k,Zad13g"Ӫ_uT!kA 1 ўF ޹6TuGݎ0QQ/3^j,'EB9LX)1ٽt@>` \ 4sXnm[gaD)QnՁ剤g_e(eIb1 ı&6VwjP]0.qj#\W&+3\5[P][cEiߘUp6B`Ѧ X r5 E[Uz2e9 îf\ㄼU!\ "U2VjNddb@fC6l}PrV[,-\J}S:n3J*T[Az@g3 d2}m v-~R] ɸMACXW Hh wMr8RED&X˼84vOQEg-Nå5g#2PUX#ZavT pBe7 :)ؙ@sB:%&.[:*]{Yi T gvJQ!]`(ƭEWP])FRl mՆ8uKw3J1 N"wW %I[|Ϩ`+/oeKgQtY#y:4&%؈?/V$1PM41Ud#8x4iȠ ]m@wv3RUYlcLHQbcUa9iF9A|PKȾs#vׯeҗ6[_ V=V>;w!#+MY +In dy:p<\K߷` ܕLF JW=$V?]G"1,< Σ&:XhhWh3‹;kN(DMk VC!`uĘ}ih^BB&Ȝ×Me :YnmG-ۀ8rÏweS Uߨ"bΩF d-ЗBةu&?D^n=9uP'KSqE,l2VbM=%DAԜS2Ů}(TKK }q̡c(!Gmw} ((\LES{X1e r["@-I)a:y@ 1^ABqk]5pە30cmGjz WLI"rS."}]iFĪAEEA9)j~'+U_;4;;ʕ-T`=댆0dJ̨I `L{PuXoߪ'ͰlCV 7;6b]f@_X%כ bZ0z@) c-*1"@Zʍ(iFb íW=u{E*t[ڨlJ%cT-،Ӳkز`fELM@:XV=ش=ɗMg&u"2KCbR{OH6'ǽ[:=k7!|!APW bT>C^@(8jUHd+Z ΀z=d乐ZfpJ$nÌW{|tw= FV6 Gm(6 qLA5 󇁷\ .: vTS.*wZn(ˡ#ZIgr)x:bzBe. ZV J#pA;t IXD;32B>x= l+7[d#^|yy} JvNћS8f2ԠwBR˝NW9=駻VY1xg:g=9jGn&~5.?^(aeY=oeG'KzyTszV/o"i|x}˫k|қwW_҆_*ѯ#H+?˶?hu%T;e'E+׷[pվG4k;PV?q:A6?{P9uo u@B.MLLLLLLLLLLLLLLLLLLLLLL<[βWn$N$~YuvF;a]P(ԑ Pg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg u$AQ'r uة8PGau-.NQc z)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙBg#z;-*ɋ.d'\~N?Jk-?\\ܠÃpaevL^y~L܀.1$` `QF#95X6^El0(# v="7uH ]` E ֠ωj֪vd7<XGA2v%u~`ԅ=Q`Ke)Z?Xm]TÀu]-|,V敓W6r<͒+J,XO}U!grkZlK# ` d ` `xdVl (aI6ּTVdh{6n$ Ku6 !gRudn.Ȍ(Pk̐áH CQN{h4FwTjB;BiRM+Xb-pKY`Kb 5Y&2%QjB%\3_$VB,)SB B^R*@bZզ0(0],.+* 7GZ\x6ʫ&d6"Tg筿Jgk,77:߬a^2߽Tv-AHf cSzj %KGk:<uq8R]E0| g%ꙕ^2GG%?3d%{ RWF/F]%rɋQW /_Q])&hmU'-jP6?uTrf$G&ӛ*YyF]roʹ+~Lo޼0%C/V7rønu"PxQ_v3^0`nF2r.Z}y4 xR·Z~>TӀD} r4MGŞX@}1j %]$ MGv&( {$H`8s?܌נW"в0H?oQIj-!X7^~%~_VV]>w[[/ b] ՓG@(R ->wڴx*>,lc43~D p FqI+psauOU ̢z%Ger] dn9O3_aVQ>pO J;nR(ގa{/A XШ8Vn4 N4UIWiB ° ŚˠNb-t?޴%L (gAs"p@F ԊhҫW »uXiWbTL#ƘC J"x :*ӷat}82&cI';Pl=S˃DC .#"DFr4gK)K?F$_A5UJB{P6zP{rQGe'zbǧwcikDߋnVLho ڇF^b0uU4Rv!3: cLha%в͊\8#h4Ci޾ۋ^q6AwNsA剖e/=u܉ni$d2?6ÕPeiޏ4ւ/K]ue{a>=Cc'0Nczv()i>Lr!iUM/P Y UcmR(B=6"RF j[ yRU8I1t!2=܃3Sl?YY31; ~D'r@YtqT=q1|KE׏W~]mj{ ,~8Y0&ɨl ~|[~6>/bt(FRLSY6g}݅|IvMc:`"zܗ" 9,8t'w'SvIN1]WVBU{` hS]IR2T!:cspizUvԮ| ⍪Tc?.>MWhz fBvJ04O1e2Vk9Em\p-bz" ӿq(–,Us;k\i58 qϓX^ֽl6vÆbR!㪇' on~cn6/5h][=1b_OuMqo.-ӛ6͡}9x AB^ X~mlyx =/2&AtԿŶTzqK8kքx]{DL,)Swu`s_hpDHӮOELe4eZ`*I㮾CPC"hM@hu:N5?*h+.ڊ)np9No&MmN}=7oƓq{SRH?jl<vg$2_zލk#rv8_ k4%65uG; y> fm&`e|vS729m{;ݍjG/*(Xeq~WNpaY(h4 q+|24I?0`,ra (kF2Մamc{7?k,td2oZ&USV/AzonPi|r0e<|9z5Eh%t$1\YF҃HdV Fdm#5 =FBTrW|U6W.Lޘx~ZIAr,2e@;X<GyYN;էw,5p*2@Ήq9w0OЎ>,byxGӵhqY^Yhx\Gn􊑠qaTPdS(BhtԜF-wd*[<:.Skv3UIpA< d#0 HLibjxWj|2Z+. smz])רv( QF3p8 #|z F&H8j aegɰLfTʗ@Uo~g (X+ u#CLTLyi :kGtGW+p"[>Og&L6ydpu߮@8ME&hgoqw bXcZ;~;GVYvy_] 7~LF.20͎SN9ʜ5xm}Fwg^hN͌fghE fۍlu<%0T毙n O]_=k{4T}Uտ Er$] `S(#WT9 &_ȌIl),GfğǁIc,b%ˋGArUpYY_ ^ 6og ?z7sDa6A>7ՋJBh:2HSZf}SfB^#-;HI6dNi&b'WF!ILG12pZH`N%T G܎&59o( gWBY= a A]ᰶИ}-6w4U+[ ]_˲swgtqОf^l~r#=$⑧_) 9NFqhlldXuVsDda(3=FlNۛSxͩ];5vP2ZԋWz1x2Y2hz&iSt]yig`\9OL3PQ/ &S0o3#X ^*-eQY{xU3&H`Q8=N9E`kGGHA7h> ,,d,aGpZ`DiaG t/l>0G;rnzcyioa f`p: DyvMTvTxWn系E+kWuX`1NcdAKǵDmK4l17+s{'l@|fF)PѶ`]59hEɆoD&"RiibjxWsb*OZ>$ W)Soa15+S{ yF`Tl IJSp6CFޕKyFCȮ=gU=JÁ`f#+@yWNK9MZMIVeåC [+?@5F Xհidwe's+Bk UeA<9VxDy޷c-wd9+q<5!rta"rgܯ>Aft+wN:br7Skֻ3)Rrss>j2#Ci.N!!w7/WBBXxZ2d A4h Ӿ=w9'O!ypwҔXx4=fVY;7]rruvVˎ>jzPt:L}㷔F@RJC ;hߎdweup9Go8u. X Y ΄&AXBE+͹YFYx7!J#/+/ ;4RQD8!Id4rFӵ,T9_g%l`?l{4oJK^LM x+<,bhHJ*vh mmޑ˳9^ ;]_8?#*j̎>&t%tU0QJ Nq [zb裩Z]( +ܔ}R07`-)g,N מRvl8dxWq:{lBB#p3!@VJ <+rGgyVS g,@.N>*$8YqHR\jUvD2pKY]Yޣ}H.|NDi[.F9 63Tp >pIa'|"A#}yY-ʌߎgfc-UjXQr`lp1<2p ,,'=.CEDsRh+kcp,_\Ѿ ޕR=*Qяwp>MZQEWEe °6j8,rITˊ?B|Q*뺍{?nsܣ~p9 w3oi8a!Z khϙg6+8,5Q}sTˑ^H.?~>Bv4<{ayyekmd ,HpwAPO[Dzd{}h%V`C2"YG^(bVRo"tYD,.D>Tb,#|X雵ۧčg"ms)Ҳ"'sv]TX3sPx߲Mߧdl3O[k5"l-WrGP8~ƽ\kQEKH~oq~a`6Y:/ }@j`QV*mKx9aۨR.Yg%jPy$!_CPC.@WhH n P5 zy kqȤrhIuQy6Cϴ6ACU:bҧ j}>3l (o Uqݺ~1 q9F601U|ACq=Y ŴGTVX4NwrIoGK(!zjfbyiW|Iw}5 w=N6ϘaCF`&TŪϔ&CF}I6~~v1 0LOex F'_0ڤ:/e44GT9{4!YxBU9nU=ꠐCFi8e E#7lDw˨xs*KkO}Z-W֘mEUܹh6eOֱL fܚ#h= A|jn^^Wxsidq>dp a˗V#k=m|>񒀚kc.}|o(ؘy·s}fWH(5x"3}ȇ*|Ql,zZg~֧29sYkڌ+ƭ1c˽03ZZnO|/|-HMU/+Bgz9@vTj\B>AɈ;Mn\ܷW߼F툅 .sjm#:k2]D]8 P$!樐f:.qL'UqEÁDg31iyaԘ<5&PSf%S__*hB1-!ZLDv""[9v8D  bŲ!xAn8og7Ydؙ~8 E <أ~S ^k31 Y!esk󕴰Ft0OWH^!W({5F X:d4=GPK9砘y5q⅒?!PNÔ=2k.E׿wΗwrożU:1kS 9Nme鐢w//O}nCFDbU礵&z'+஬%ߟ;wxP*?H䈕j.e!qm"!rl^@S4.)oH5kչ=f!nر W>0"1~X+DijL9rDj;V@kL*Z8_T|p5MG 3 sm,90^ќzY6YYP RUUFVG vg\>b߇_̐-> w 3bf5@)G8ZdȧŎ(X7P5"Q6eZ}Ң>jBg.􁭰ՎqBnCApvլȨteS璀z"xuˇO9t#JQpy94īq"kuKpW1!LW`w}-&okݹqîЫ"|kk>Hw}"dw}u>?*u=׻tR;$4N2ibK2)|>G Gag5lKGiuBRiJ'xaYTq/hwȇ\YQDZ^RR8.dˡf&:]t¸ m 5s,<0URAg^l<\&hmf29v3#ڹk٦s>]s=A9(VclQ4X+ZP|qr j8ڰl2 焍 /Γqx7|xn֯OK*Q" =x_:R+̯)%dBy9Q0q["< s+eCFnoT9ܟS..oq5K(6pv<~轷yGSΩ3҄T ɨWe <>7kHڇ"VӅ_2YڠmìaXnii)d:\~&Kͼ!Tڌ?D;aNt\/꿴s!)l}5ڌ5@z + ![#0:]ѿu@^ CF8%8컋r{pq0Hf\=J=|DBH. ?G<+D~,-> l'?ڢznRIpChrZ$ *-:>ҳOz W3(7HsQߥ0l+f( w4bÑޢ OJǸ6Y|K4-z޾.Dפ Yl(Ml?;>%tv:Cʮ` b&JMgA$|0A yjom*&}qn!0DCH|pNi$ev-Tss#eR&\?ќXaux'^cقTF/.QWFK\[ 2=>Tߚ-gʊyBj]zvm\p-b^-?$Ў{DmAȏz9zgM)0b9ĭY!!)xʹB2.򗼋+ھ֖8JMT^q @+Nރ4 Dmeهe q"ɨapQh&;@Vy+;,`zH !8p`g/-EZ8J$*{Zn^DX\`= ET tl.hK&Qlb1:(Ԗ+Bd4V5=Uvf[S#iZLN,X]0muAg(OZǧy\6"zi@]ֻ-r7$Ȑ0H[X0вgNH)ID{A7w- ±` $  D\ VSJ\Xt#@yWNKyPmA5wd_cBDq\7ò" P!qnz[t µ1 :HHLu>9 c+uw\ A,`GAk50r)g:+,)PA!k٪c X7p>zCǕ/ɽAgv;!Q-#$<6^+#x/B8oQ!ĕƨ903NAkiY §|<=-3olo8p *;_b<6~TV1XNTYsR:' !2[ːʞ+p~]Ul"*zu\՜k %*V·:}6Ә~$8QMטD w_+}&T~4p,A/seOq5e{1~9d@3!G GwFxaЀT`;X>_z k',֚5>9s\ޟw-er'6>FʝץZqPG䘢3J DRH9ƿu7!S1r c3QJ >IT8_O(vȈ-il>_+`~own}S_zfܬX3+;{sY-yx;>_rgY/U}.` dg-`ͳ!v"ü|V+՛+:dG\'WL[_t[ߴο]YwAnuySX&W}Ƿ"y+Y틩u^uY/ïß?Ϋmv͚3য়w'TQ5P7) MOQf'M~/t1keϖtlc`3]6m7|Ӱqn)ןn60J7Ļ^ЎWջLҷS)7~>\? _Wl5%}V-xV<,gy<2x&5&)(K jA2Sà0࠱VoCVI%bq^#d=0pOoفm\kwce'յsoN])clYOqh!˗/\$(7B")NJ5ӹs3kkά N#^ay C*YaBpߕJlm ~1zW!`c^̜[dUS~}!etC{4"ɕn"jYtշ/$Q )01N\]Ygk]s8Sd^t09mˈ v5~nŘL)U}B` #QJ"2o*m^ײdBF+M{% bT8f9Ft|TI6R#F&īɁ:zi,P$Ѱ@x_oJ<\$Ɍ8Ļ$H2#L<.uv !G9tB|TcRL wtyU3;B9"4JU1/|e Y'E48iM{{4MQD5J2 UM~_B!k(ʛ.{5G :?!JKw\)BtR/2kcҗƠA4cu{~L및, \J+ϖ>%,oO.[+B[uVv+lq;.K;/߇׹\VI~ݲJw7R[CP&.GZU?o`7WN@  }J |:Tr4~K:P;#:Ifӯ>:l.kɹ|)~Ny$f <2TJ5S{Pz5ג+$9̼%ެpoZID8F$7,tLVދہ.!tH6nܟP"#9$g4ENH,N%X~h*p:oN0}xĔQ9C_#I1h=VO:zhҌqD=42o~W J&H],{~/ʸu{+o e aTUvQ݌ ,ՐU誩>)\5X(B#fCF 1 υ| bcG5FɓUJیCPy3٠S3k=9%$%Qo H3C jtu圸7}J 9H _ (T[ȈS vI4yhќ'&hS.M pЦ $ xCQ48ːɘ߇>."7mq52-ܹʳeC9AT D@})=g(``$02 'e)g}.էe\KݕGblŘb]%"N#}B'I"gc u Xz~ >')%4oVOzAM)+Kt^ۙ_freg|0l^"nvcUeu z\5:7of?nO=yM177A7;?t-w{Z3/eYQvMXMrS3vX\NZRoZ+c&E̦i'j[:a"IljLLw? gG!ħ<;Sxq4QgaAS;Mou_WZ돦7>=Rb:|ux]MM<]a߅J#DDZ%y.2;2 V܊h[q1?=Jl>BD WI82a&NQ~791_mVYYjঽ#*"Xr+.Qc;Yo\r+ yж j@iwi҄ =u{jo9CWȭ8;v/R|aտʚ_֔JD+6_c;+yomwb /u0*EN #*>k)׋zK^x7Hsr]o#7WABd0aa>drn,j[,)=3A+v5z̶$7*Ǫb*rJ-bUC4GBI\L)B7'[bmw9M N.13G>{ws{z8ea! >^(TT3B +5ꫣZwG~&Ljf_˞S:!\gs-s}qD&-+^0l)nG91o[]r8|ȋZl$'U ZYe7J=irſRurA'up1@&ӌb>$ *O aRݪYv~U=zO|d7ޣ^*܁L#[߮荾@ۯ֋ =3}/(J aLo/ zr3aLB2P??>_toR)c>3B q2RhsTƲvu|Nzh8K]?s@ҍDbk"<"[ vaW5t7q(3BjXi0l6] %.p B+v'i4#9Y(F,A| ^zDa)!!+b%!T-'7mP~a p`TΚ=8 5J vySZ|n`yJ b-LU i # P7Lxz}hl#xcwM:3bSԣz߶}u<ݍɱ\Hm$A)}8\ A hx&h?٭s:Oޔ+Ӹk:Owc|TɉFhv N>-(@އLޅ=v;>ij5]u0JZH!bVs$ jP00BY h۶Av~PLXv6LjQn>-%r?xDuT Rbcr"4;bߔ.MMs]sL.M71c12IL)* 2KF n쮼~ncӸk;pRr>3׆h"cĸS9 S- {нn]өѝ[)Xw,<$hWK6aJ umz&5Z̧;SO+sP!S08R\itHl>kCueۗ><6oᣟ^U/5jUHFNi|v|=jωRCByA .ʂT:,цh+:T7;*9IWX!TB\1xACf<҂C0_Y<8 #JK.0= \Zlh~ ΚGbp|ux0t냉GLg P\2*0 -_ʇmXv᳗ۍ_ȌH/s .hEvZDXR/ƥV!v5uֲO81 x1 ۰PGG71rsԍ(+>8HrS?~ vL0~мX^+J0mgJcvc?`u9MtȥagRD1''EEGZRP(ss`#JU"bL=-̸:ylEDL8q rfp*L)T(Qo OPP z!9( $EKR] sA39ȻwW7r#18 ^TnOO ʓU6#kDzF{xkK*Imq5 n|~׾Z)/Q;‰ N'~Vƅ&(c5ă/|>NW&ħnȦEA U݄%Bp@$ zS ݒB3Dy;׾-*1T9FvQO ;m NʉuϨ3`g`qN=|#18`!>.վ%d>޾1OH ;_( Q ߔpmb8uH 䘨TE 4{XxPܨ\lp,sXF҃-=:Jiw\a#-8C(y#18F6No '!)"0y(+Jou5n G/ aN;x$<./xciN~=$VoML"b#'F@a[~ ( p)},>]M<|>|)鑎B'lĄGxA)P.^ǚGbpomjsNs9FqA]4GBI\gy$G Ny;?PYR⾰baޞU:&IbxF>Jfp(G.rkTnr<ְB D3J2jy(lG71,I«C.1*x\1âGuADDN?+*i/ ¨"gݐs;>P6= °Á4:TpiO\۴<3+hBFoI py=.QGl{D" Hp游#18tQv|'11(<3f-5w~ C(z/{bZcms!8hNKqD 5$FqE8"YU v1k֫hԃ9܈R©-v$ʀXiS<}æL]3&k$%Atwr4ˍ|?k'!` 1`E՗Y\{>&U.crsD!?t_0AA@o88!_LVԆL[lO3v4ߔD _d\z44P9mg|vBxW*H*~hՒjGLN۞]i!Knio[ *?z.LA࿖ZxEI%Fa$7jcaZb(I^:KP Å|;3vgE Ls]4ן-[Pޠ_խ5^z>j6}p0`Dn{;֭:0ݟ! ^Ճlj.\M;0v6maϗe(^rv;_֯-^./ 嗄lYޓ~=H,ǞapRY_2$!zf@͋\JF)K;w.cjzOsumO 3[_W$=Ajib'|'5.E\ /] 孴2W:,Kwv"R;U]իquKmJ,ame7Gu?ڀ`r,3Ź1B΅pUN-bJʔg֗ 9 8~WٿYn*,+kN u[N=Z Jha iK5壎}Ա)I0悁 UrAr'-"gd{NDR&r-X6uZ~ԴQRjڟZ(EwNڼOq乀ixijibs".wm]լ4kJ#fs `T]gk_ [Xl\֌&q^ݡˀɛՓ0]8PِξBuX^_]*.1\++5 9@ȁg]d7`> 4|Zzw^~D"Vg7@"M ;0j>s\ <7nv]\]"}eL{@%zow)Eaȁ 1DeXT~ !¡UMM̑šq#z{UC- x(P]-ugM^uݙye$~&pLƠv!% ^Q tz`1$xkw}3jt~RUk2v6}0Xr}ُ?v)(SewxV+p'^g_v|u4惀QDrEI6+Pj %{Ɂ奈_.Et/\iQqڿfQtgq79pHHpE}C@LO^=3Ce˲dmIzZTUEփT(!Ul=P)m)gIэV ڣɿ9UwJ';R:c= OP͑m ?u דE`?,^|w#!bE58xl;& nc`,g$<đ˺0#Β8уsy7HTo^+~j?[@,ibCf| Ti /[ITQ8sq v2Tg`i|;R4^P&eRzV,<{2ep뎧p fa&^&p\KwSq cR[Dc̢濫e-TC;$>'[蒋 \LjIӸ%*{vk7_Ҵ2{02&4 3$}m9V\_O&NErw.^U'8?ELڰotO&V eTٕie~DPj6Q.W%Dt5gǺ-W%gw'g9DYEJKKĦ12%8T2'H%{XR+uvő44^ 0q+ۇ,Q:X(]۩:fR`6RඍTE>ui K +N@8C\f`bw:W1ru e)w\L|hT»vK%U 8J/ޙhxf 9Թ̣xcd*i &Aơ`$67lVVym;S1|gēuU1WewA>% al4 $ESNEݾ2%ʴs[ 6H1.d ׂ ϬyhATb`9X~ F% 6/E:DHasb}.[BD:l䫳lG׹͗YD/.f0.ꕯz=.?fol(»{],'Y_O`i!\`%Aw_`d/9(76vI;UIМ%}4yP_o:䲝MngojAKp]e&n7Zy [nsŔPFQ//ɱ/Rq)d?n}&LLqeS`4Gya <at`*?)HQ]%MfIhQk;$Ky5Rx0"3&p@&vRF mRR V_#\J& b8AD51H"\ 9eX Ò1-a26lp߄y(2~c#٣TVOGlxӸSrt;e)c8 t0 RJQ۳RWJ$ -F!7Km2{J\%Kķ\E Hꅢ1GDsi$3@hkz-ǎoI[\m7០c_t9%εxeK9X:k4ʹ3s#X \}qJpb15+5gDjndU ȩ$h{$$Nf69qN$|Iqwø c:-]'j{ X(F`8 O ػt/a=:Γ@%eKP!=_6L22ep+j8ſ R[@C. cR'SsA0xf-(= c,(%g;FOSzgY@?my4V`F&B1ņ W(Y4&$u+{ǚbe0< J{:vgͺuY[a'2Fk OME|W t>&z;]ۅtw*F]r뮛yy0x?ŅxfAعgf뇜^A-aで`mp :!gej&=^a]yYva aBΈBϘS=A ƈ/\@;4]t]S6̔ s6cjV|YrLUNP!_Пc(d +c\O7 d\h8D]-Ǣm7]Ysgʹݒh'$NZ" _G|~^. .?dJa]KSkR5{x0P#"B{ Zʒ9?oOJש]Z\zbaq3wJbj 8Ws{ǔ90LkEq a@yЊ-D)xm?ūF cfCǝ&.ȡD|ǎ;jsҧ]yhi8R@54hO3hU%_6Yf}?rZ+fT1ҿ[&4@?Ï򳋕R~Mw|Ij+ėN'#<d>T-8Oeq jNNV /T&KK)Y֠$aFIMtoJ$ܬ/SO1Lv}}4}JeF( Ȣha{(Z*-n۪I@@Sخby jc\M=-8"} YۙA:ټ.2ٷ[p}`vWгDW@!}hϓ]XE2S@ p|IqIGq`k!4\ԋ:펫w뾦\Al9(i4!kA5EtB3h:n"3}3A/uLЌ21[NSoFcX-l4#l3rr:\oҥPiF>J^ǂk:- uY=]¤Usk2=(Tp;#>cnɟןsо{{5| dhd)AnmK#EV* S7C;OB@ZtğWב8^X{RgN~\ę"b"s{K60֏.K ڗcP?>/7o>B1K-S$];E;L75kc44%+a1髀GM޲Ζ)[}I?HB 13Dgb' Wգ$ 8g2OO\BRuQmxazل`}@t:q$t9T~o@: R[@)=;lmHk3ho;oO.8FC/>ܰoRIL?B/U-c fSVURϻ5\JV<ms8O~yU,cXZ.8(ʾ-@Eo&hMP"㋷ɓÌ%}of[Weuj+}dڳ?n<,o#)V~r|>>F`jbOWRO8VQʰ%&g8pP\/|G( u.Mʺ{̺ )Ź6&Н\mv3=a_,锔V}g#gtLJxܳO:ѐhy${{[9:uJi$tbYPp> M&yy)8,E_.Z6oT7 ػ7TK"yKfi*&_ V_~f~|" nNbIK:;Sᩳm rMX$8ͅEXAs؞.\QO.58ۇ(s6vS$}D4b)h~2`)#BۯIt~nb.X&cwt"m OGA6tEpB)C@']`;Q3KT2vf<.iw=0(O>߮'sBitvɵ Q e'hGKukch0q-}.ce31,rTk>yS/\=KLR\.oX̼ ܺ8e#Aq(5)<`]t`N;(o;Vxά]z ^cDdOq Mɫyv0!(:Z0fd2^@oPn[no8]T#?y"Rkr4ȿg}aمbvqfE=V:).%Ck grC*kf Ft5Vt>-P6ﲻ^_nvMm;svqg ˬP. D~lh$Vz؏&W}%SQkhW XxZ{_VfY h`Hz!-CպF2XІ974gNt:HF38#Ӻ*Q &PMT93 Qͱq!Dor7X29L `ab2 r,0pF|a2Ѕh;AybΣv[9xS9iՖ 1Nb48gae vMNӏ=tHCBASspc8U4H,{JQ`Ilӂ aU2 A鎊-Z&t =Yu]4- ,S OiP@:kc`3/~H<."aT:Uy@")W(wPi3y̒34+^'F+mc`H9 !\JSwT\M}OéSh2ض50z@sJLK]qp& zayÆ?J}Ni5Pٿ3ާSRjc`(?kxR}eP~sga64\K 8Ee)krōɩ#.45¼*Jb߫zÅ_g&av1ӕ62yG 1آ r IR=""c q&z]M}\D1[DRV1 Q뾡t+s(N7`o{ԏtS#ܥׁ@Ho"xg@+҃m4{wpto=00GvS5AWuuMCҶ^GZ*wRRRN,1"2$hINX[~)RbVr> 1m/u P9ޔiXm LOP/kSn<3mr/'󑯾NgcgmO2 7wE‘B?VfCtRؐbVr&С~ Iam!j[46u1}d>x0Yqpf9oB6}Ōgaro0,:/هp8,2gBW@럖]̦-=aWI+@fYj8i@L^XnYɒ͇G MQzFFD'7w-R懪ʞr>~OI|ʕXE ްs+Ic.. #z!@:Fg"`"RS-a$)H8H/SM,_$Hg{-`&f9*k\٬fUEYsƳ۱9}vܥisUU.Ob)#\ڊ f]v mwr|_*m֦ M5LbM~uz43vh|M s-k5G8/nOͼTYe'9͓,Yydcfh`V)Q"!@%KŪh6A7Id@7{k;14*҂w}FUs==ܴg-DO.+D[(D'|<ܰOBWLX'82Q:N4qS Jw^}$JwMj/ڷ<'(}=I>x\.V8q'6\'Nnӏơߪ S-q$%Ԕm}J%qSUW03·O{fԈv^W6ít!2i83LZމ0A謪ىx;j4QURI<ͧ{rǑUa&чvrҾ}5[b_4N/I"oO|$  u8x9Tj=iS k5sY?yDѾ_mis?x8I0[NHW˽^MJhdP13u"Ό݇`NjI狋MQ|9W l < stߵ7y@!~=S1]EpN{vlvK}omNyd֭w?oKiøYprY/EЀVB燶ˤ#$XX'K4|}ѯ%JĺjTiMmH*8]ŸV-B N*qX% ,=8~8rO)Pݻf"6]ׯ>h0agFJt0u ŧQ2pʏ^I,Wem}ҊA ,ĺ64] ֒,_- xwyD0u]f֏_DoVw~ټy]yS,6j6r^Lf5o߾jN:XY-Kq-r/3t^, "kT?}~+(,팮$rgs%7ih 5(§2}r|M$z>VsL(QmGs+:2,"1rbrX+/B")|`#$՛8R8=yg ۼ[ VhډB1)G|&.|=K~!Gݹ3uŪeXv%y9 VCJ^^.LɪsRvVٲZΤ ]Tt#ScEYU!/¯s`o;&yg1}ۢh5-3fD.g&75bInhi!̿},>Ѭ|Yh&''͈fنٻFn,eӘ6ÀM2;d3&6x5-KJ#A}dÒ,[%۲@*>N`ޫe oѣE!Vt #,`*OіoNއwH,otr&6zGvlRI^"c$xxػ?'@j؞e,=ܾ32|_zol^U|¿a))UR䘴}17ֳqG?!/7X,b4x~f` {S͛;&[H}3I|? &r =?p |tM06C{>67lo6loύmߍ+,[̲\`, c1=kNJeN>so<|v Yz[^F+GI)狉q!^+Kӫ]O=m},kWan{\e 6G&>t}gHs>B?GM6o6coy[Rwؽ=c1T7e)h8g|{qo:"2K`  G!=Rr'5yabw߂;1ũcUX*xcu1 gJ SSX",N{g0j,zU ޷ɼHpc/(ީNGL1| ղ9X`\ڔôN6qG|7nt&KE/^f]4sf,t.ANXRXR!}=YrEQհW[bE!<U,V!]Ynu '1beP%E%2KZl`kM|\ 3f4j|PV^14_ }I7&%_ +n!3!S۴ycz# ;ra[LhF3xy=ǧ͏aޏnf.~&l ':R-;bLA^%yAɃbAC$IsoA{xA{lg-=H^ ّReםחMct]`B<UX-zg՘IDk1 rl5ͭrQo`_r:y_vmPXF9N]`DZʩR ˽)ׅotcǽ?N>g ]p i*5Ӈz& N'o??<.Sud>Z^PG Vp).}Ԗx32x5t) qKqEHx6N;gy_nsZp88b F-LxdN0+D{-"#a:0- 9+ b}AY1 3 r_%AXGmtaX{A:FHH5pFBb%#1))DXҠ8KFbRA:aӲhZѴiԴְ(5MP6Q;]{VAG e?9{D ]o.CE=c.@ouG<:0=9Euݲec~1ϧY߄]P=MT̲PY]0<w?JFY c*h%{gqo]2u ^͕y|pzy5i }Y<_|xͧ;jz%ir)`'AË `QJS!-Hw&n0-CT' [S@뽯3$Q$ ?\8p?xyߍ2>$^O~ͥoگ\3DW+:+L, DZkCpsϘl'n | |\_(+<[gXۢHnsI5|3,;x6?t\5~ogk*?k|Oof/MioFOדQ3xA51le=(Eb8}i}_On2Oz-ד?ţ _ik:XyS+UR[1EU 4;˜,+u1в |o_%zE+A+cp89Eg[֝ *B1X_zs+ p) VGgY4UZYi*UoQ}+L}_Ni`\)o'~'" VX cVj%Q%5c`Mz>0sFhO&DhO&EBE,pHC Pa)\ )# 戓[4}^.xߗ?0HH,ȁ@Ln| ,' _ ۟FyoLT1kFW .fVF\lP6U;hii4:d5;d=5W<}L۴mue窱)5*Loi=eM&JK^iQJ' i|us7{oM]@DPÐIz3nS~}QߛqCxK߱xdKqv' BݻJyS$8\{s?G!*UO~0oHj*Q ][o\+ av楊d]O6X ٧`QȒ"ɓ onK-#5m?DtïX{e.7V#s}ħ~z=e>ͮMv ^켇YXjF[)5oW[+ v'AXc֋>ɽKcw'ǗWXT.|%r[NfM4vc,9^xǵHi:ܝ֞'y>e!a/@ѭ/nT^R|G/gqt珏56= {q^Ou[lT5dM*a",ڔP:L,'j7.}t"kC0i5HWs˛G9k|}- f_o<7-='c.?fKGSYu8;R勫nNPF}z08n!q>_s TH~z ^ɼƙ W{ݰ0plcY ;}_ӠqbJol!bK!(oKݢO/GĞR@J~anh Z=x{GOCGwLS dXXBR2#?Cw}ז18U%`T*HR 8bi #h  JkSfTuEO C>D)hMIQEPB1^7~ZD?q3HG9VܕΊXH>cޤjrA!S=|TL5h 3$+F_1:e)KFE i܊pbϑ0<ŽJ:)lE|3Y1r: Cykt5@" 5H'ϐ0֎oe`biʇPdIJWіY-' #=HTeMշ4djQ;4c,1?GhNTueE %T'Kk2 9gX'HhL)^W[RP ?c#Gv #S4ǼcI5+[9(h@)gHqIs"UUrT^A?{Q&ϐ0`E4v(M ;q^ϵN9F7Qy!a %)~Y, & #[hSZ>"ϊuD#axL2$PoU/n1}?-axDa8dX2q?T*犳2~ZpPת$%'GXLF'VxK~Q SMz.s$kS2^uArBJ"D#ah| )8*1JbQR?C1c ܫPN9ǬuolUJJN\QƑP#axn0I4VUi zCDϑ0 z{!敤oXx idd[?C $!IBIdԄ>)ʼnKJs$ [/ߥc.nl[)ԛnV|$!X^O(aԺVܬ_R/}m3:(1>qSQ8]f ub{!G'̩gU$Z:gi #Kj1$-2 j(X&m' #Gԣ{^|T9eiAE+6F1OL̑0<;\9GrǔIo$qϐ0|><:"uh!ogQy⮄ig#a0||"i&+%{*qU#21Hބ8RDB)ӥ2F0|I%V˱qG}&d3as$dL@EbaTxi #Kf1:%#gZfHF|a%I9Ls$/QsO2愢z c( `b #{|.⚪&`*N+o@UZܛ#aQ;ZUvBN{"6R1$21Gw0b3f>eq: Iج;G-+Ta-t{J+-$?CV_+ 4-m6B%P8ArH^"hZ5IZ" } C0z& QRb7YImWIn' # _Ӯ5Je{b&\\rD$e #7;jVԌ՘FNdus$ŽVؓZ5Wɰ$QԦ6Α0Hl>=}ӣkm0Z5 IůGVk}H`Uo1zk-|}uW_r]aW]Ik4z+}uW_)PUomUo~uW_Bcٲ6|%}dzſ7/ƹM464R|󝟪/~[68{qT&ɳN%gYyRW6;_4!|W,ߤQ DI[5 {@5Zb&~R&Sbb`4sXPN:+:_OBU^[ l^*֑ǫ^>_*NBtqKb)eY"#`/vAt 74irH`Xl ˅f"Ubgv%{[1NnyOE:TŅs ¥Fd3^uQ؏e-VC9Iƭ쥸@Vu/~]0G=mj=9PUiU{m_84h3cijs4^6H<4O&Ǹ1 |G`9I]@lt`8uumѪ),1[Nx+La*e%Ko'SḷJ%}u(ָl2`)E< J69Q̭&4ωjCKWzQ %7Qs=$\P>7ٷ`*+E.zI<q⋃Tϥ 'ۚ>r<\}ZLf"h,S& b{*vEAg/-H-Cf$I$YgDTtUl|kͲ/ Z3{ , jH'j+՚-* )IW'~ZrD~'Y9HLjG`ɜi(>-ԽjZyꦽx,J и`@XA4^]Hc1BBlZB`MQܯacZ88\Dka/Euno_׋Pۮ _|-G}Uۣy9_7ׯ?M/;.\۟+%eI}+_^<߬yrenr?S'ǿr#Kw} >&z%.Nqx' 1nT_\|`-|7\.~|ft]{oF*B~? v `qd@AO[gYrD~$%Q2)6eY6'ȌGUտGE/%-&rÌ-;9ϝc dh%m,e4$R,<Œ+6y|;쎝 V:]k~$0`k U c(t@H-@?9HyAIJx->Tv93FZbyz9K~uma+`*w'qjtjVIn6 S>4 KA8&Ara 9x;&;\L&I}ӮRK 715ݠJB~յI(h|} pL9B`>ÍœU˸hӅrH4TA͚Etŋ<"kq91b^,݅_ߎ*i[TUNPťP~Jaۺ¨ݯ=4>~?1Wf0uYE5Vgutz5(vqmj{GVM ?@G&c"Z2O婡8vf0H[Gqi8ʑ/ .כf#>vVa1R|3RkSzDtgHe!?HTm\C3BX.SLQ 5a AP=QJYj"(,0 "4fa')TP-q 8>Le4i%RH`-R~ra J}E'iJ$:;^q$y>a6aAۀO6yУ'w 盔11:=V)2aʽIZø.: '(xA(_M F7[ylirAy{ɝVhPzq tF=_ DhD.ָAWS9&4PmGs\BtdXDcM6"VD h$]G}05˄vLg)>Ȳd DpJSʩ:DIz+[OԸQRKzT0ϾzM\1IQFi*ktcro&zGj>| iTu?êaz >N᪾C*hDWHopj_N' ᤡ%|3W x$ )RIJ S띱Vc&ye4zl5BZ"8]M+L泇.0Ӆd/L|ͭߣC\]$m& Ym;v̋6aZw'v-O>GԻ$0Ϡ 絞)e.8<]?=OooshtêTel “4?DƞU;Ynt}G]x no|67 dϼ5dzV~kGԵ\kz6rK^4H,L+SAT+b9+zEJ#+<ͬqlƨy9O rJ,DDꥦ0+Ct=ҝ~>-AapwP_tWP?;c;L,c05%a4UTsc%bDbρ.ONc? ot[lk{w$ֺs DȒBeJfǥz3̤e}ngM=/-Qkf =]+9^um@]U3*4)9IEdQmDafXLN+!Ё$d'= J٨Rպ" uQa3C-}Z8hEQ*u)*A(+Ed4ۏ(?L\O5H4l3Pd&8FPm$3Zi@:dԺe65@pcH4,02G$4;',yb1uYM MF+?ޢ͂m4uiD~46g@ɓ#I8eɜ(Cu˕e2tb_.2-:y0>* ?qmn~s&[nh+o [D]$F%ݟoA ^tL`*~Ӡ@X =fD.{&wSG$I.:: di!|+6矋>, T[iS2Xm 4O'h36rEb1}V pʷ[]04WA| ;(SQrbwt5MTaL^"?9O|Naڍ/W}7}eҗ(oƫ,]W,%J-esZ żRa9q~0Y.L?MQˇ-^.v">[X|/S&>|~)tzczʢJm-^RtU\eUKv<*:A,UdpĬ(Icͭ"k8[(_M/:`m5~, mߖ}~$~A%Y!k_t`|gH+>MNn'N[ɷ뽤x;Ww cULh@UhuRGΘ3A*NAI}1ږѤY>E+8&)T"BDD{ǎ:s1a\bH_er>`\TPa~{BmzѤ3X3L&w'7[޻gy?CzNj^zdm :c5(X8;o^ȸž]żZT6Je!٠LfŃ+OꂞN?3KFB΅~ŀ]f wEΙW۹G xug|䌿 XHX!>TLy]NG-oG}yX#Usuz7;_ͳR.׫I0īqY|iLD'کM&cxɍ~ܼd[W.9ql0O1WNJwuI[jװydwwpQjNvv!OW=nnTUz)܆ tϫ4 1w[FŒ*[m֟G b=z]A.BKZ&h{T!]6j19%T686{V\4d*6ԉ.I[qtq}Y (Eᘷ#I/2a[C7]y{Ǟ}ˑ]6t 1B)KY_)nq]+qсK9)np0sHg[E‰uXTRGAw3Sp5A0ɵ \ 3f4j|PO^1Tk%i|z>11ŝx]ڽ*WpLuZ},-Lc%IbI*;bi2PHPSQPڸKLJsq*+Gm$^Oj51ӘG"h@QATDGR :d Fs>N-{σ޴:H+eGRʄ&_ߎyOf| qTyv{>bSMPYt ҖG1z]Y)\Ŭ" ̒;%t<"偅)x|ߐn$+l, Ǿ;IL1oO<Oǔ!D!e!x0k5f,`ZFL&XFs+%3s쟴K糫޷j-2C ר]QQ8@ҒPNbXXH.u;m?|P]jTnnʀv4imrO~R/Gɦ%wgJF(y|oGm ()#32x5T Nxl=w=!k3M-@Z<Q>e@!A|4 j-$* %^F@HHLBDN.dBˁ~}GY1 )fe  D9{mH>r M'G#L,=/dH'r)Y(WC*cRb2Sм:=_ 6knVjƐT#"B21&5@ AG!,;~x#wg785fe{Jj+/U^Pt=c+_%_T?x65{uL+i8I2[dԣ@ k "N$b۞y;a Mr TR|haF]+ۀo{8D Z5'lښI$'c.z:q@bĩb$>cL`3 ^mqz'P*=ioGe//f¾ v r3SGW=IQrDM4`ٚ鮮j0𴊉isT.ޯmE z5eSՍIo ~ "7'HTWՑxs' FcV KV6YA郳}V6<{ O1gPчG߅^Msp_~nza?p/VB[A)b"L/ Y%/{Yaξ{xP YinVMo;7mK V"y/j_-+dC%e/Gחِ71BomI)|{0@YgcfE{gEv1%Pcq4AMgH Ye=^vj9{@ uEE`RGSJmΌ6>hDՓk{_=1llJa^iGjd^ wE%{@Uᶛ dvݺK7ho U'H!%g3LSxKйG!,D+L}.eQtNԁNvC|~]En >QcvKeBNg+5J ﺻH 7VN7&B۫UslUsbu^o갍ؼSv+n݄;=/wQyfzHjn/[} o6{Voĺz#KN6Ҭ_7ݞzQgx}ᚍJ'V#Xxn'H9cJCKD}l߰|ֻfZY3< V0r{}ѹYT!bs+Zj0Ηo'.:QEz2aDD&p TQ͍QmQ'($8UedО\#0\\sp׼ Eׁ!y"5 4g@TF*ms0(nWg'W̖]C_9<>0sjpw:%P⤒[ǙiD$AH FHةH8u&h':=%E:Rb&y`D¨UQFYJy>%y ^#:>% ӜAB>-Es'H<`~2]I!6ZWFW r; VlBQp5(~i1ɰwJ5 WHFFH<Zm90I`%Z+.8OHFǤ"9̰4N+!H&yݗ ½ P^[ NzYېAKbB4"(HttR =7`+Ed4Ǟ(?8t'jT #3PLr,YBh A\BZG`aHL;1RtbFpcR(B:HI윰Xs 6F(b^uwp]/?DUY{&΀;2IIBhLW,EoSX>ꛧ.W^omn.`|.U:?Nxmn^;>y ]$` ⯲Ί-]#^ֿ861&yg2^I1 528 ߁՜Œ,uwtC44X6_iEü];}&:DVC 4zŶjb1݅b ]Lo.a <9wZ P,b>'ĒbE*ՔH3)F2<\.nY6H /_Q^?}=_2U֟Lfa|U>_No,%Jp-uw^$/,5l8zyr7--/v">zX*.>|D~Kl0Xgaa=k\/@^eqgU 6)g=VYfӻPKT|F^YQ$k]Rn dSH_ƅ8RU~IM̦~~X;vXM$k엥`fbA/bW\er @?w i!GSlXNitpug _ﶔ߱7{~EQo X'uo8SZ|t sr>IXJɼ`(㽱nE˅xA1Nu `\ڜ:9fOm Ig4qk2IC||}ShԵ1þ@8oZtZcWbZ`gvѰ~>} lPg*H䌴@\ \Ј$rm˃w4*yW?B5{YI"G7?N+z X^$K{Hφ`ef /C)z-3t5sM>mU5#v3msz2\,'#ok>^0K7'y|%T8 yӱOڧX .XZOp4%ƑNn?1*7*޴K ݭ`-hJb3D]5I 5gwPoė|Y'\Ù'<1sų>qh_]NLbcNg-VL $V{uv*ԽTjfsnȎm2.tn*~rIʭl$nW[{WjeY%mB́A[o '[p3 fWzY5<0HnmX"k@Pj;yCÄs3h1ocJd&EI= ޴ӭ]X''kW\Ccb)6P598y;8`oNYБ@2P%A{8aUR; d?928\o!6vI!w =C;4:^JW<y7jP{D`wHϷ ~=.$Kn$]׸OwSz &GS9H.PHx3jg>O5iZ+IdZݲz@0{WV7!c 7Bې*j)K7T{ >&yZP \o+g̻w?vϧ^Lw%09"9z0ɓd3IzB,Iem m˗)[=z14]tVսN|]dI!8k»l% jH?/tge˳3q)wz4Z#߃8a/p~.ݱ=5-0_TKI,ieps+-/*V}0)#%C!r{ĈX`LX[E< J*':=u9#sO${P15=Oavs3+kAm1?p=XM̘N0%5yU`,!ҙ\: 3);y-Xhm 4wFmɢQ1,%! jCc z|OƦwP8l)} ~i#HaP|3M,49 I/aNt͙A6]Fr2}&*=s.f3),Ln&x9) Ȣ`;XHH!tq*=Slto- =#]Oޑ =.%!'75̠ UL&߁9Goq΂f>og֭=PF%a{l!D!e!x0k5f,`ZFL&Z i,F?[{q=}wɬ Msn$VOb` F PLPx %w[! #+~ X;YK \0)3H"ˋ߯zf)R"[gUkDHG@Ϳh>CQ 5FNѾCQ"gad;vsRP+돇W:UREw<9.K\#'=o k``6뼗Zs$wIC┡V@)Zc|(X[J؄7a_)o`CSuP]Z}X&'4;dW7q  ٷ,.$LP }{Mbh$c"^&tRʾ KB*Ft= )D%L{/XģOMH؝ VZ{A2f4%D*@5 $( Xʂ*OTŭs[}2fVZI{JI;j {؀Ya*jcTuԇ]^[&/j9♊-$JL-}~A?[0?vN\ 7Ih NSy:brک$d^M\2F;SF!ҌE Qd5MXaq]u>Cnu%$+@24#O?8r)_ƝsdږEN;{Ν=YyYv!)eCEMSv*9>Trɿ+>*!V Z>o9%MAv#>+cƙswoNl'3·_u.ϾT WnkāG׳FtZ/qjmQؖd)Q%^AVb'.+?EOR`ޮZ9k,&I#u&R-/g02hO0ߩ`.ʱq<'E-HnHE2 )=B+1d bԌp\`)˨jQpAE8{/A` T9:Ên{D.b[C=zAQBGц ʀtA-LQYZmDx(ԣ)5[rM+3 ^<*ym7KU-V\l?%%/bTnX(Qh]%)miDdr*ytQ^C?AFpLo+AB0٥ Z8k'yu-!7,zkKjw%_z 3D䌴T:A# K \YgRJƧ.xIȘ\mH8zyLب`_;]'v5D`O ]{)F?aH/zc!PW@%wH,F (̊hXƯ/e x1GZS z 4q~-߼3&H|2.\:Zῖa~=U3 I{MiR=K(qxEU5}[Ԧ (-7yp(u"ɨO@G,`PJ@`xRKQ/gkA%O"f7 M3̃ՎSė2.{Dp>+H#1 U o6F Z/ ąb"V'NuMWZexڽ= .:0;Ĵsc:2}XDj)~eoK󱧍ס?bjAuzq1>j v9%ϩG#|(8J$#FmdFY[\ " XRO0d1FD ӒGct[F >ldd<39#kł~!u9GE67 #[bE磍.|#D={D0A?sD0r|ވa*<èpX|@D,ۈK33HqWr;#nej|j@FZ.X$3+vWDž8K0̟HjOAE?"NR\K5XށH)Å<Jf>}ﬕKn{3^]_D0j7* >lדV1z [6}Z6]ЕSN2sW506"<P;$]]Vvћ.Vc;-@_;#*!RD OLwb"2~ALi GbS6Gb#c 3Tzdm-،`U&XQW\EE]ejT]e*l70΢RB422s*S)iAu\h0G`CѨL.Ǣ2LTֻ&Օ&T))p۾i0Ӛ9Y×u}ruz"-'&/  >(L+kSy{;9ec$jŕB $t8FB&1@mNSHd?%#5K2&D ZlrBRp+@﹈,nSL! ,r Q 0K=ILjt#ߐr&8MP4 \JBKDN)FCs|NSPj2k$TW-v'#h_ Dv9oیnwCaǟ{me!#U.7hy#vEkcfHPM ;|ІfНigR-EO4mMr5Lh):Lx ۰7Ff ^g`4>;Zj;u{uLQý>hkO's0M\OC8F=8,&(^`X e uW9h1$^2/u#9T'dhQC^f#^;{]08X\5Q>LC5ʹ6LWZS6l:JL.RIܐl J"0tb-hoϢ]#Rf}$)7.K\e9%"C]<)2<7Q`f"[klg^i`En8W:zaHh.]r%S^es54%y.4JQ#e*p+V;E!2aeeh h89[&f|~(~}T^v Ȕt}D+=EK}Ey.6?:Gս5˲D{wmR9cW! #vଵ$heԨ l9tIsG.)I=;ۃ)%x4SUR oTDD%ỲACd|B,VXu.b0*$K8،+}b(&yeP*ԿpnE2,q-? g|e:>f3k&@P7V5Yr.B Q )ʝoYAZHo 1G'2,Y 1I%YքR北F`Iarp찥lbUjeR cDܔ{&b82*!PCTzy\-.i2s8.ҮN;AW~7JgĬNC]K{Oye;~NʽL.?dA< 5Ϩ3:jPGL C^R\)6"9;m!&{w&ʿT2񻒆}_fmuU Sc [aQ&ƞp\|c) *찹RA J&r]v .coTXrTԯ*s_Q?~H7{g9 ^][Unóߦ_Fqxͫb=GZɲnufz2·_eɯn?9j0DŖL;3_4,3, 匬Z^$tz!6Slo: g=>ojѭWYYj)̲r?NT)ճTܻ NO?CsƭkW·t+&njُ97|=~]-w+okk _г+\݌Q X~}\Rk>ﻛCB-nD\Ss-b]δy_,Pt 60P'&`,2I`h#-P 1UQfۀo;'~%R|z~D 넔 9(^>6% :rPG9ghlj$A5wm!C7M㉅MݖI3ϗWey]T54ˎ- ]t.]WS:*=B ]i׍/8"X " -DT; a[Mn'ݟ)薠etY+o=BTeܴrsrC$.}][o#r+KU_,0yI`U+/ZTIIC#ڙf꺘\x+p $) d'DU[|uϼ{xꛩo˔pX_MlhшTaLe/P/g>MU&#nd6ly<\nwG&/,[3y߰Lt1m`}w>"Q[j߰*&=}*ANA7:~7M˝[s7MKmpXqGRUI4ac-4@|ĽvV#WNs9ܓ.`kcJ|Ix@av]QјgQ)i7EG/Zl/x4 6JD褡-runƳOC@ oO|o@.O;Od0ARB;dvEU"Q,ݠ׆c6Xz d?929&B)%OI[ u{v3C{M'N}ٚP# nG w5+#e5٥.͛32}ԗ=c` ;08V$Am62pHCYAgsT#-I%DV%Hdkk,8.4*Q%fhڬb1uL%z3ޱBʎl-'S7wthzznGt ZQw\0c)m:J=@;AJ_XIA0jV.A7^u5s= vKL{6!Uzߤz].ikHviaU,c笱HmR!8,<%g ,/| )0h-lO7o4%Xŗd7(Ie1IVˋ$41jJ,pC 4 Eq1c!eAQJ(b@rcBe-0c}ꋜ"/']5N'_Y[Ob_|ԓl5gˠ ^&C jc(̡\ʒU6$L)e6jB >G }d+1gA댠+ R)-d0dXjgkP2@IZ&iAKbG,SQ;A't쨩#*)FrP`+/݃Wٗɟbr!E`k {łi 5[ցX"G))ǍƹܗRd 5$(9[KaZZ/k'_c."%N9'oA'iHւ(&"'QhR[J'@ !L;.x"sJ&͟p <ߴ2EiŮRz6HBɗj4Fgf4ُ=8^GX)21X)3Pn6N?J}qǢOu D$@ 1 x }<UH% }& <+A|Y ʯ70z93P%82fl .$@V\)R (Oίa63fNkսa~C{g|0?RaC@nilW:/+K]_fF$7Tbə#wgCޟHW?Z+I?Tsn'yJ|o?og5H9 G1`"XZ=ZPRZuD%rR`j0z\dQhЃk(- %b2,B1{r-8]riou)NV'׊}/S޺yi/atp¯SNS<(^}ZT )`8(_BN*\6ks=;zXq9ł%U1 ǥ+qJwJ%bloPp`U*K}N\]UjN]]U*54ujԕDeϳC0pGVW"wYE4GBWr$ԕy2M]= +s3uUv(RԩJ%4uՕԵU%؉.-AE]Uj& ggxd0g(6CqcJܸ" Դ]]j.d:[3ZIN\̉߳B;IdLߣj٥Q-Ţ^,"qCn/|we:?e??{Wȭ俊0}(^ v/طA`XId[ݒlS=m`w$bJgAC۔B7y: yS"2k{w}k?ݡ1 {Z(#uES L(Ul9ETG'qp-m!&{7 ¿4Y g߮5 vqTjT]>E߷v8Ƀ(# \oWCm]Xʟbێۋꥊ*QsL\̆o^+Vp&|Ć+6jc+/?K7~Q߭~ϛ|5lKn ~nKcϋ~\Mwqn U5=_Rk%ki)NDg>s~*OwhMdE?yd˦ E=y+ulzӏw0ιFuٷ5*voyC[R;M}- _,Pt 60P'&.2I@Nw(@i *Zpyrl;aWm95Rs6KOh} TEqvwYXI5f4!L2>yޝ={T)<}N㋥T ŋ5/(R+){L%}b(~ ЮVC  Uj]Qm$Z(B ZId$ףg{Os.9[-"Z %1΁eB@\AƉ$gA{ZwL'&gEX+ŷG6:eR!/USNph&(f+sQ*#}< SѥC_if2@evZ 2 č4<5ݻjw>0!RL_s]K'M9_'胢o!]=>_>+ב3mċm?ؾnT:{o)ɧKu "z,*H\@D hf ;N +q&e%*G go|j=+7ydAty?MEkbwV؝sҼ}}'|٫p=_ve{|3{tYW'f)[|,6<0Im yCUp™ΰM؏yWʴY&myȼ==)λx) V6LD tңv{hؾ,+HWᘷᤀm+OQᝏ M!2*,`< _3/wo`[k OE&xNL~ >6NqS7\ݞLeYWX}D%u!@jvۿF\=ġmumc1 bX2w{?Բvz;T?6Y߶w~ I'j~wؙd4m[|-nb ]:ŋo_wY@<8S/ czoK r|rmtrVCA*c,jBj:|jԦD8 8H[X9i|:d5VV:NR@ V% A0FH(i3ʵ'EYI@uNbx 0fB,H HUaa1rv [ 9ZsM-鶚G?}rzl ǫ!tqFqiY8 \ITDeYP=T=Sjr)$t`Inuز–wmxl6q=ܾrw;,y5W-5l;q[y#YabS7vٱ4?ofZos\'!'[knn~"$3P|nADc>B]=!z7\e3hi͸\t4D]PNE^^d =0.;B:X&9eSḴIƪZ+ c/.cuD ;RAk<$2bFXp\ ou($/|9 :TVeh{D*Ȯ\UXHeHn҂XW[9-g|T3E(vsu3!PKn#PtJ*P3+$Δ\b8{ ҡVow+Z[//:en.*+cH$`pWNASg%AD%`,$ETF9$򕵎9"7*6x F2xddJ <(x`!i(U"#DEgI{9+}y6>,ޱ##[eD2Ii-zfx␔BHG1cPvp:4c`#{ĺ k.&D(R騋h@G'5pQ 7eXi."?Cn:l.ѴU搲j)ϜBh,YBh A\BZGa HL)/!]X8P"r@FQFH`Ś;o1B,<;ȽUw7Vz 6l GWb{|&΀92N~!GTXtEov~~L>$GM<JN{7W{g'vCiw?럓f}OvWDo2Qq>XL^(#f:o3"g3[КD$M:9: EMRLn2'_{jvaPy=m3bj3[|8A;{_S0騒.Hͣ#,rnVnr*fk/jJW!`/BwϹj/-j&ӽ/(jg__Go~}7af%U9]L/,%Jp͙tQ'??M_7`%&3,4l0a55 u cSEE>|?[pxm| mumx /n e5e{uP<՜x%Kq: S.UuA.5wpPr~ج.ʌWUNAϿ\ _N( Bof^ɯj¡]gs·s~W~<`:k&Wb xϐl}8u J-\)yla풄mޕa@޶6_Mmx7OC ~T}]u͛7t)o5|3 5=Ĭ:C t_@e'ݰUkjje`63tie$XLjAF{;-`_(l3 mfG y1tu pWp6˖1 " x[^W v`W'vd/[X Ytg*v/I.Pe7εQBMC `nCX 2l*-C6Ǥh^k'D+EOڷQSѸVNK^W0ҵ)$ O]cbr-A6U:~H?-&v :M$u_,*sP͟0S*(ֺ7*Xx+Z}^ `^uZskC )nETxxS6y&\2BV2!N?-e {[ntϹSfgȿU&>ۭ0OIR&>-s[lZ<$/zE :;BMKB$wOyu`n#'Q 1|!H* cvEf~z32C:wp )޼f ̹PQ6'r^p[yaٻFy6O'޴DW:Wf/x/Wrfe.2IM)EbD5f(+^%7یWIK"9}]rHȡQ$9}nS<ݹ?>tGfPj^'S-x[-]-|m$6 W=.|x;X(rVRE].!^ňK( ҁ ADD߅҂EH4O9Q?EWda1|z[ӯ)y} ^&wufٖM'v__RF {eO˜Y8FYettst%T8 y*_Ov8%TbhJ쇍#Yʓld@ ʷc܇?Cw{H Zl>T4`͢-`cvtJQ #eQh ]:xfX[DS Jo+r$J7MQ"5ܳ' ?bmhi xQ2L:.wL\nqUtXe~I%9}Sjڔ7$oۘ1`=OGkfԈ~ΞCxCoSn[-3{JLS'lPj3zϲ6l 툷1sk4Q㪋$n@Sq Xuhɫ*-4I}=Ujvn*CcluEo&I+NuYБ@2P%A{8aUR9t LpcJy 7|)m;&#ͨ6=:\ E d]Rч]%RPOH { VR>W_K_ӯbsqy3h?SMPe,hV*0RˈaG鴟Η)7pMfFʩR`#ZH̆il&p1$rGIQ:; \OLFw^+iԢE7'' _vqSuub砱㛎:J(mISbC:cŢ6qlWu|昕.X@JL}(g+bB-dRbJ6e]՝"sElm@j^Pj$Z[D(B&d&Zg1Ezb4~'[%n͘~smƜVۓ_=RE.yvR:4KHs-HiQK̈́y@#,ho/fz*Pb+ńGξ9}#A*}TTc\ fJɀ + 6W8eė q^*l!ثBt6Nw/qt9HD+蕁DZ1$3[_JF 3J#|G]L O.&|f`S΍!Z- 8@xss4VYE>%:,fLߌF{ ^OePzZхvKgvldL*wY7#J,`u?f_8v֖R2N)bxЄٯ/CKa0[}!B>+[cR;ڪH Ԃ^K_j#ٿ2 Cnz85>%)R!)J%"šH#24]]wWW)Oc.*5bmͅ-k.,WPii*f!JQڪTR4lRU?XWI@9Ч)!})~u`NhF#,E#\~36%  A+nUTS(A( diQXgAggUQ(R<xP$ H ?)U> +`ۖ"\ɲ<%8J^ot^ٛQY!㲲"ިmV7aT2S {uVQ&(2h;I^M⌟d-0 )0d_:>~7n xx%߅M$S`̲[]0<RzH ƕP#! <ØGDb 8GJ0U E=J @TS B70KL7/˫⛙׫˯2~ew}Oex)ۣ'8ʘW`J{Vi=׊ǀ샾 ?堟TVejEÓ`)\}eBD(e߽wo$Og7QH,\+s`2 X@Ί|+tQ~]}h/X1iuH.Kj vNPg,b9Hns@]v7<@el`83ν_t*-jfJɔ  Q>A. QE?J<*!TIZΧ=^SQIfa7@_>NHL +O.VJH//˳0'<\;_` gCflI4ug] 5Ej$"3)siD/6}TZ z !OG 6#Q@>g̐TDImNoӍ6z$JDC`2# a3KձF9obv[8t>ǡpdVף<ǟě~#zjY+}Z+L9:kky4@XXu9Vkn:AS>w{{۞LqB"(7J]&Cnf Ȇ85qMOfYtzug^\L61 WqL ίltw BgS \)5TׯCE_M5*46GU~c,:u~?m$^[6A_KEY̙C:9u+j1".7kJR?RGSJmΌ6>v;;Øڝ.|Vo:wȶ> 6`MT Z.cbo80[m`b~]%:vXUCgW ڲo](9&vhUWcaW Z]P]}j)="vE5Ex ®+X-v3~D*,Ѱ:c;]%(1n7Ȯ8Bxƶ:6]W҄+d߽߾="ALsPPy,h+L<"1l&ClX8& E$JH0H!,(63#:Xj$:Tm`VE59"H}uk5s>R$e6YD.Tbge9WXrM$IX̵Q*g er+gI1i崅\[k3b)%Z[J,\`>U Ap+Ǹ^>Y=F=ITJ Oׯƽ'ϧ=7 Mn6RYuoDmxodQ>Ƽ*H/ (C€Z;P9;/9u۪%e~p3=S, i\/ C\YSٝJ/|M-_ av"aG=S_%F%7.M]OkW|!v5ۭΘYaB~u=փZ ftAJmg"7K*%L#Ø X ҂ʦ-V$IrR=fK#˺%t|τ1\LKPz]ՄI!6M]00@6$mqFw(\܂ tm"anI%nGp(װTX0#‹P#U$\Y;4sfZ#հ"efp U;\@U{@gƯGU)*eT]$3+q5`ƸӬ4#Mh'XZZՂl:P*-jԞ8xr4riֻ#q;:RZj$ht]', ^@[c4Ie\k>b_ƩR7b3ח?x Lgcvߠt 3,X )nujP{"cnN6 ˫m]g6^Xbn^6!ߍNMLryå O=cF@A_F1Tk%i9Sy3\+.,- Yon.B?C$"o@&PD'Y65,rL˶ܛOM`s̻=O'z/Ə<vLz`ؤa/[fzϽoXgX)eLoU=zKռwiPjԝXx^j# xȈ(q)W>Ut<0kMI'bm1C-+GHPE/IM y͔;u*Na=VAIDn̝t38߯JqKP+d滨S{ܷB4g/BN`Ƥp)--Tɫˬe8)s4̚-4*kc&PfTet8rD`p')1p{8.6;Zjk982XbS ,dY*vƉFaz}~р5%7i93 H"WY[krT) S0OP0rqƸ7_KP5J#"PB`bQrAQmGsRȰ, :EDbD hl:1pV!Zq)xw;Z~vOY 'G ̚W}jK y~[ ((|YX<6Ri+=;s0S};  "E"vumC_ 9/1r=}+ m-1 rM%N* p䬐*8I"2mp6ZFNc"ԙ#ap) zƒm\t 5q _Z9bL2읲bo1+=AEV$0 ÃGIC]`:[TxX Rr"0ixJb,"B,>æA)=[ԙ)Jା:/l&F(R騋h@G'5pQ s%Vh ďm/?۟J4CH4p)2%TɌրE/y*FiԺ%\H9ƤoQt8819aN#ZMpE1>ۏeD~] >g%G^*A;szZ->/cخ$>B5rtmy0>J])\4WI|L ].^e꿶~6x11;bIaC3h P,tmiSV]U b?t~\t]w6e]<ݵ`%JQ:6 5*:ɠj׽l~==lV ɽ\d z|h\׽V~nfz7JVֽjC1z}Y2j/.‒K2L.}HG!M hZ~N7_)yasͮ# xcUG1ڃ-`ԑ3fLiaSy |O9K.!zO<ǎHpc/(ީNGLR>EY>,ڴTPTnp1usL{ԥSv`t|iD¤S<=zUX,(E,VcZ̏eCҌ%a >@Ŧ:NjuuH6?`x7ѧ,Glnii<}r;h-uvP)΂( ɝU6ZY!QtNˠγ*G pAL }Z5)Z޶G"OY܇w&9vyfoGץg3+;?޼-rJUʔz^$Pa_Yr1)-QO2 WgìKMT{y_n|Wo.g6bH1`-Y,W<<"1ř6 C1ݾtiEڷnpu-<_/s 6<"܅ֿ^eR!I.x$ )[X1c`G佖豉hj4BZ"ػr:K^VQᅍ&z97cA1m0aT(jS(DOD0"j][[7+~rb^x1`C v66E:%G؃[ԥ[jPyH>>~ɪX@UPֿ(doҨi"RVx]9]4Ϣx :qYɬ'Oc'z8m̙Ӣ׷=tpp-wVic{*&|o8L>PBzYe$d "P<[9Sd R;J(je-CZrh(dr@4:H T SBcAɶ܌JjU?v՜XWj4h;8R_;풜;PڐlB,RQ8Idmz01* y&2% cMw^H7l6݃2:^I(moxw+q?H{6 sgԡU Ծ+EQs&ǎ@1cfҥ'ZT"L$ed`TEmEVJdP_/]I/9d|&uZιlv Z* Do4Q%0[ }$@tN})ŧR RRZP_&2u-*-XK_v-ቺn_uˆܺf և^Ij#l]i8y5{!5[4PNpZ&mbu Au#vi3%jN{uz4$kt'QdB(3:` Lj'p(ۭݭ! ѧrM6t,SN?OzzzM5)(&XގUx;.J_|݋ d1[PWޓ4%HBtg7ŁIm!w?)j h1_Ԣ*C5b,Hos'Ӣ4ݹ|ӫ^GΪNxV.a~'OKѧ 6Q;NLkLN=:B7T3ī%]&^ qE2VWώP~0ƧXLvuϚ~D,8H"ݶ T|MJ,BAX1aJj".'jdGu?I}M14RbmS"[x鮙9njT"n=R9FMq[Ϟ歷ߞs8qӛ6%*n e-)~m՝orxF[t}懌%(^swܘB~7ލOĝRfh.7naE;"(IfʭGN9w Z ڗq hCwwاҶ9pE}\;ŝ0o);-iaR9Sb,m +a̓cO?Է3x`%,Wh*^EI*Ct(1NOQq|?p>Vl#9;`ga ܡp`|f1mϴôPo[r9NWe0Fco .f6.gsɿ}9og0C᮸Kzy;wO\R˖O?>Z'~]ҁY)Wbf-TblX3wp\/h _3O,>]4]d|--gqvb }?/;f6]oz/)X*p2?|8'.ߕ]/x4^|}ߘZs3./_/ɯ Di(F \Uk-:\U+W3zf '9U5ظj;v*!NRpb?Q[GǙ+qqjuzNA=\=KMsjWZO{J)P`-U5WùUv+mχj=\}p5X jWllચ+ʞ:\U+u/EgWlvUUgWZN7 W(X9w6s\Y_^\n7W"ǮPd"|"dR!\^ysI^!51l KiVSV1 pT ű79] W>Z@aha 11ҍ}a?us+~ ]/?K1ޑTX?wEyLKdQ'/)tO:.T#9d1]F F4QRDouԔL RP,Ӆʋ|v;䳳1?`&}] W4z|-TGHqq+y6>9^>-3߆?Z3܍'PɎwD<3i&BV"2:فI XǴ |N*)b&)hi 2I&Jfu)caUAYT)CdJvHZRm]<ĒH1 &<8YX+P_s>BbO<+^;IB9D(%x6s4 yGB"qbaWK-\6H0HFfF,(Xld< Mc, k隧jtM ϓ4L6wAѧ*3wYh(Ġ1d,e*\(hAlYL"b mq p7TH Ҫ¦66*Oż ))k]j9;".OSAmӣv`xúdMQL܊DDVٌ( J D$sÂƆ`P褳e,QYd82F Ģ:a3svLԏ>֦^ "6OEDm="n݆o[I!AK_L6fA3br@iuv)V`$$hʜ݈xg:)u6%OE]=.4d= JSx5i[TsDWkɇ^z%I {K6jqPr."uV3R(eJ~Nc3O:]|~Y߳Y2"텕zBRq0,rJΎH*(MWM" IZ8<ՌLh2T("J%BU CRW[1&2.qCQnUpS>v}I<1.I΁N (mH6`Wh)($6d=<mw&λ/w-fVnI&;%`hAhgʼn5Vs~/i92͆:wG sgԡU Ծ+EQs&ǎ@u5S)YJ˖O"L$ed`TEmEVJdPSG6Esdﯯ^jFvg;) '5\7A -#sچK)>҂|@'50 VyU}E׫VF?i@Uc2 uAu#vZdL=UjOUOdc!,H+O޵$̺/ 3~ʊeJGzz!)i(R%͞z|]]ŀG.Ȯ"ń2;udWcb_*Bv|N(L1PD* yA?a4H/}mF~ HJEVn9mꍸƙw| ^c |pSfpek~MN +RӍ45ȗqz'k8xKӝx-V)M?yc/a-% ;J8%X I{nJ JNh-Sq"`NZ%HFަ>x&y|t~wXvaD;.J3=ǝ8k7WWwuƼICc7&):keC& V4\uzRyz#던g739*SEFH%Z}7:f %5cfŒrgS S,#+]wi)8nS'<hωO=kRCU?<g5e`YG B&tlJwM>mlV669s$cC%G7N?g.KDA2XE護2딅r q\'F΁K>?ү9^;­\\7v(/>>,m#vėՏ+~}scY<zwcPM[LPN1"CH!ߕ< *-Cfkl?Ylmr+km|zRjٶ.)I뛛,J-=a&xzv׹߀=xW8[e`*kU>` _Yi=6 -uAY|et3rqjVdkkƭu6jVX\>^7 -fyM}K7<>WN mm}=L1G㊿JyFB&R-tϐVj8 kT#.zO)Cg!pRbKI/6']/qEY=chSS0wN3-ٴ9*HJDҒmV_Zs5]6 7g)7P:D&x+ ]q $V^@J7+ Qh|%l _W8,;&[+ZhhF7E&;51'|,| #-$sd9só$C^Xxfؙx;>zav)GԃLj߰*'pQ$p"{}*F_w-wNZnc |tnxf=v3Hu:7G{Jos#tOsn3 LҼCTʷ&&kt&<^/VC]t`*JDd@_E_[rujY6P5N'=m7yO8C̤]) 3"䢉2AEA{'jKnxNLa ?{ʽc)o=o^Wu O^:zL7jϥaz]ek}hk_5+*d gWwwt}M6cUC7X*kJel/oapLn&vb(:}pAfv9`􎧋lp)?+힕=8uGQϹh6V\64#;h|3J'EGLK'@!ߦ`>YBzR2lP }I?PaT K>kmds%$8>٧}RKr&,>592,w9' )b1'\`əeIܗNmНbjtNC9|PO tR)JԦ(93DQ$so%!ia2!47c/CDG"E+1xϴIF@`fr$>3/)!d]Cmbz$!M";PT0 Ȑ`!f,4JIJDW'1f" fsa-8 ."6O aϬ3嘣G*~>^Yc}"*vu(@K!:K-RL$pIRa9"s1Jq9aJ ͊t ƣ5 $uȾ>at\̮w:;;,TY0s-$_z,HwQZ*a>@n%5(cX29$ ә_I.r^Bc1S' rƋ  3 ^Lʰ.)a2q|-)VNvkiGД}K&0<8$2h9RU݈*AVYl`DSIW:թ, "1AX1y6d0t ]B[ګ6 SKHrW4:%a>C RzYfA0#Ef20J5Lammʤ|y1ț0*AD $b"B):k$ 5Y],*I&3US9Ac81l``&䚞Shmђ>m%K ie^fT<O ?318* ǐIl% ѽ  ìP Dd)UO tRKj' ]i ~-Xep  uuVr )7c@0qPseظNbՀ : N&lZ"X^!ep$ uKkԯ=΋x Z@2։ፀA8v d2P2hEvUȝbfx)F_#IwmIe~dwol[`WgԒTloU3IFꪧI 6}rې>.܍ugO̼66Dnpp0 pcGtɍNŠqb&.%|l Pl(A] H_&APhwBAn3CD~=c qP st<5t4)V!hNaBɡ%P&= nLK2c_X6TgP#`&yYdQ<@@~$0bwojTE.' 0k/3,(Vƚp"$UԀL()j f:# Ùxf jTRp0 E ֺKI- v/8",t ߠnHtd# yI4i%G/`H46Z: k`d][q1H(~U5ǴX2mnP`kC\颅5k&{&cvF'`ʃ`( Y ,:0> E1)UJCz[PuHaZ2P?ڰ bR0e+9Y62 L#5<c Y*|3 Jt1C"1چ ̵ %??-n<5\.A¨B;P0"m002 RbYV2Ah}_J/\50%np,!Jq|s,5Z]>%تPT]j t;fJ@ZD$l B,9SKjK1y1 J6`mܳ͘ ?Ɗݛoqlܢg't+s`Ajԩ+}}?ʢ ^`Ti+xityOк2m5bMF\9N^\riX}'-n}8ڎEx|^ǺXQOQ{}}k'RQmذm~1m:kPm wRVv1ULYSՎReuPDiPPD UT*2WrUߚlC VއWr(ݹ Ebݧlpu/RccW:J>hgs"|dR;epI kV9k-Ftt!}\+4s..y}sGg8G 9dz<-r?YOᠿ >zM꾩otqCjG/yS@E@eZ}S Y@OOYGL^]>0I[VUn v<q> ytu/ XWOC#BT܏޺q2ߨK2D w# lVL*_W#| /P.{•k$\*M+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"p+pR xW ; W58t W %D WZWD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\WD"pE+"\ J`VF/N^|Ő9 oxsZQDz@7;~[ᇁ<{~a_?[+jǏ\A\JBi۸>\~ )yT{Gc0^GþGiųGW(bd抟0vcL*67. λdʹ~S3@"_/% T7r<q4B/6[e4b9WY*4/Oɉ!/Yd\i:Tr~1){~]&yN`V?UbWVJTUC}y;Zy^aΪqNÙXY?|r:m4u6ߎng]Ao߿~GV 򮮜OBک/) Cf릙}ægYQeJ"sUUEdY֌,?IvhCY͢R!A\;M!^t=E©*ga*re4U`,k;a2ľJ=a}gik!iG:4^>SqD LBq;sJJ \ `\F1W(4bPZ-B)kI+4WՒ#2WlWGcP\[i W\R <"s;\hJs7W(VdBs[fql%*_F?Ð2g95x&tPxSW QxbAssN::-r:1_L"c 0LF`|/C=d!A40]]G })*څ'LjwN2ؘoj $|{$1#ߥ<k|e 0o<Aãu9N|v PgYj,ɛO3)t't{97O F8]N_ap(X>gic;u8SwKy|{gᾣ ڥ]kitvxAS`eܔVآ^}ot>Oi-霥i]ynlL/',o8@Hrz9]p/S@N};7oG%;ȫnѭ^,E5s5^\~^mnUN_gm/֝_l|b bFq>U ZW_QAU zvv[b~_OyŨ;w=8QṏH܄%w $ iifi]Ee])_*\֖XY.SeLIaiNS?g!1z9 C^ P{ȷ{rv~k+,|1f`^ab5ͽJ4V@&ט:H_D$+Y#WȣR˭RC@!xO,5`{xhm (vG^Q B<&۷ ?_}P 7KgۋkGgg-b5k{zqk34;;iVhm:?ە:cv~з6\7nMu~6]|:W8N]lT v%K.ꗜնtU|]*{f{:`QlDy͕U<8V?C#v* և8"5^̓=*yc;B"W"P؄=cĪ.NѮBA  uxWW?U"ZWe<Շղ^Ѯv[0JQf-/%ߤymyt 2\0{ʻvo_hX\w_ۼS,͍Қ9?mj}XOV?Mg_죝;7a-]lT~X-%xs>55IMfㇻLưlؾ3o`E$lZsfKS:!<++լkA(u'G瓓w`fw˹uդϧssT+V[].cwmHWzl5ИYLc˾ HX=\ES-,ɖm.["3'Ȉ3:@> ??pFLtKq}Y#׭ok\7?=Fm{im@βQP<@mBR,qYx ӞQ`h<1[w蕓 %}~?Փ^UL+T}XD1ڤxh%`90)|cFU I޼rMzI1hф=oɤ;W7ųg)km. y_r%+ ^reyOTk1)yNi!eJs.$뻠xJFrѪK^#$SU^}er0L~Fh  trfYʪyXr5&a2Tǡ;xةK&bD v׮k~sh8?<<4j (EMڎI]̓q*@,1 I3"9tD" tNXkjހj5`t,[izc\DșAr#::!<.3qmVJ#S[T*EPfd[yOݡǂ?*lp_XjEM `t` BJ O3M!ID"QcMtd0ĪW9Ns>I69U:!1B'ʑ\͜$z~I?Xbe۩W'IM6k_o䴯 $ak ̾?fkG.8KAϒHNZ0)f ZLÃ"۪ȨɈ9Vε%=j*>ax&ZzL U7`a+X]^saz[\ԍ4OMhKSyk !PP 0a7,e +3m(GdBR!h@#L. Mt)hl͋ZlGp=¥TP8 4EHkLVqA$D),@ցDlD%zD#w)HoaV9EG a &HFMđ0^B)NΞfv(a⸬" {p~(_/pI]2 #8kG;``Y'%8!@h&9Cs*;goDb~5ٚDS`{#,m۪͑7+m{VRQhG3@ȧ5@ 1+gJ.g0:SHkbTNTEMt&G@ޫ3A %]T<{5LWR9d|'z)ݥ> Oax0V(o\iuh sv1gdXZX% g]I) qUuݦѶ[[qK]&A>Į^,73 [X#Nz ع@?>?),~ӧ_蕿fl R{;N4_^gv! *+ղN_i&ɲ-"M s#sC>8 ?i 8P8!cHGG:O}+?wCsvo˟btM=6~2{O} `?Ώ'A,`G zvl :Bȑ3OR1+M+WP)P̡8U|/™" :+dIUQ+vmbp.4^q.4"-ozt׹>e>3~l]]w(`1ȭCdi]uk3u|1F [Wغn]5w9tfwW[VrThn-ww{r;=G-͞nm_y+l;qk%^'}+Y?ZOϚn|jS_=lA70.\'"RknE7!U\Ry+eR z9^jDw1/~Z e/uͶZgO{Mn<4p}^# 6+$ö+"fI~O\/|=ʸ{iCs&KNI >v`萖2 Py?b:ěr[<@Z?|x"lExzzC\oT="zuf*?Ngk5O%_߯ϑadY7g7TA҇*&hNcQ3A)Tm82GY@g^dA;6.O=쬣j՛G Ljxa>jVcR|߅3pfa˜cͺpzwb*LL0 xD $ű]zґ5ly'?ȎN.d\.A2S9Ixt8WMt9W᳜fTx}!_;v;Lgp;߷,lyHcm"3lw8l]5& x/{!κx2t <.IA_vfgh_~,|/͠Iq8vd86'{#hTF: M!2*,`< _3/w`[k iPOFbMx89&|ݞKhТqg6:؄M!(ɫ|EU*TN8S%&rmlTMnzLX>mS7Sz 6JP %&2` ܙ!"Ox qCxzzɒimƦ5m|&3lFC|7$ay$jm9A'|zz*Oyc5rjk͊NGw+acGLFr œ Z$6 + JB3M(%m *$&SP,%pK]1PAG5J*B:jԌyHzUaR[e|R炳{RaN.'q0^61M,::XtD *I62{+opmy:oHN;kd4Jc,o5E,3J1&*yb;Ά79Tmxj3ږhN7ڬWD%Evs;,4%+ udP‚73W`,cI!rPJ%?@C^,M3;^("X/BB 6D'Nr j$^IzR蜰: km։pg=? ??}N}A=s3#L(Y:=`I"!Ab(u![]m=G30ssI yDUxPgf{y=6#SxkZpg^ix "W:zaH.z\$ďNLpoO\h&6F  V;ĖQȄQI ulg?s z)1+zB1+~ Sg%AD%`,N*wB|M2B]38;#y 5ȔW//߭r ? ;O7JKfж^ƍ_rcu">\Զ<gw C1x1 Gi}M NFf|qx9*X#ػʖL{3߬}j̗|c2Pa.kSmSpeӈ@aD?8΋ F# _=ZXضt|7}QE J&?7~V+ pbw|1GHQ׾z73mDyC}wwlTfٙ¯K -oxdR$7ni 6v+ö- {|. 9 dk8R5_3]Ko4ryŬ~ZFWYj1 B)ҡ76$_[oPoÂ?< _}-MEdwH ><̈́6ٷ˷OoA?Jٺ! /D(I:(ZPp^$lbG'W&-'a#%}8GəqHa2$"xkFٛQ}1[.:::!u2s,;(kvL_܎oZtZ/i|:Ю>׍"X+"*Y"*NK|Bx$lHb:zE8)Rg-ךROmZ ? g˖ !meŝAGpOߤ< :: :9:jK54flD8ˏחx x]兴(S1^"C"T>(Aſ^ O,lgOEOBRL;z}H) TcPAwD{s79klh0Ե |U Pope$ Q 8}793T{ Le*~3&0<ģ !d(AdŽ䉧P*M-:Q-Q%JD(}Dvk XQ#h1.1%Ƹc\bK W1.%Ƹ/1%Ƹc\bKq1.1%Ƹc\bKq1. _1.1傪c\<ɞIS(1şc\bKq1.AG2z>C'0t_CK:/|χzkIkzW{X2R$cb <Vhq6d!)GxEdxԭ;x8q@$Sp$e8PHfEy:鉶61Vl8{v<> /?}#4v/n?+û v3 M_KP+ܼO2 #IՊ+n5!Z#Jo6eDpi%"`8` TF}- @Χ@\" N:AH  qC 40b hf G sr&ew*G*hgbj}i*:f5OsE9g$;-ryO'~Aoj-w}iվ`ˤ3oK&BD3fs@M EtfQ $A[-ui<)3O93=E5ɃU`hpIJIGcQc )$$%8Et"y2^~j+(@ǻKW9WXM'•\J!")ŴщgZwHvzRօ.X`AX33Iw_$ےc]l,i G])Uw,-VKƶ,CJG.pNJPMhEg=CC,P۳ ?cR#x5)0IW~ftHz>+/ eZQMH.{<8ftEyKZM/_ή6+Zmfetg1ǻ;:{jO#/qvcZ/2=?dL>LyW3RL|1E S<[v58rzyyƳ4/Dt9ZGY32aJ 0MVk,ۜ|%2׌#Nv?f^w5.]Tc1K )1ˢ&y傌%BlP}^T/*&/*W{ugoε~n)z)_~*UyA^LfR*m 2fEtVAsz"Yl[!s970)Tt_IO&ya|jE)O2meOeiC+[W|75|nA`mW{o+ĵ0p7ECK IoPsT?ZHp2azDyEz !.EeO!걭ÏPDPtGהG>To tx| A=^0H3֍a~8^NZZ+SVBUj#< V-i)#D_NY䀷tgO"6b$;o,f(&ʡ*U A'CBY0Pʑ }A="i*(t{ V.bYb2!㒷$JUm95Ixvss[q_{u#w|dŧX8;ѳJ-0ӨC,eAkJ'C4:CyQ8NNLNwJR&Qc)A[ΉQmCնfMMG^6/tmO"kk!:g\ή~l8gϕj9 Ns4ٖ| h>DUJ&{L8ݗ3T@Sn?B&Tucϵڳ7gxWr1[PE -VA\2K3ܙdMog?}Ѷ4Zିe]"nroUbEN`zоpF'dښ)*5TbRFvkU3i2bWc9WfR|Zm􈪥BFF J[ڵ&ݖ2F)'$㥶wnG/߽/"[Rѷ'?iu_h2>~p]p>uTZ+Lb b|K 5lM]%bh+4p(e5lZ{*b5}Mbgm9-6Lsy.Vt^.V\F@s8jɊ,^m4j(gEg-:(C. xɖ &)&&/QlTcP\]VIr췇 s1taEɌJ%_|hASNd۔")` rk_BeZ)@|j>h gN,P[bJp xRl-gEȫf;ҚsuvElhGK⫓E:;:WP_d8Q\Qi )+?csݤ\!ab[-Ƚwq ~ye1D!ZƩR>?쟿CVdCT>Pl ^+Qθֲ^%hd8Q٢ DV*Ey;j) @F0 aY"~"xwdG:l;7UY,:6@NJ! ,G.dV"ڬ9 xfى|-GyWf%a}b{E6P:SvoNߠQs qmiZbHIyYj5y}!˽}Jd9S*CME_A\?\E pl68qT 26HPU,a?HJCVs>T{g~߿8 p#<Ĕ:uSfr;hRiCk-@\[E#P7vkEJѼ'_z6 c6fII1蛟7&u,ٯ,?3|*q(: #+̱||\gɦ*}zD5?r#1ty.}MV]pF3w1W[z8:'R&gurU;d"`kSr*q k;= ,4vNeq\VmÏPDPOe_zMhm 3}~ 0MLנkz`<0f 3b1l4;L3OYE TjV-i)#D_NY:9@CZ,Pd獅٘ $T9TBcbȠ2sO B4oQa ZшBŋ'w%bqDž@5v2m+o;uCYC.C#vڐlFtk (V4mx|[LG?N(o5ƺd1CUbpZHQDä}`.@:a&hiT-y;_@d8CU k&f tJ5|lMe)MF3>7sO%L?;3y`rZof ߟ{3$wd]oGWecFCs n9`^2cTHʊ~3Ç$D5%Z`ӏqG pJY3m3}ӕ|rV ]=;=;$<Γ[{M>-_?j/#pcȷտڟ&[}v7_UY_MzHG{D w:_)#fz%-LZ ȚsCfAAu#mO)c">kojbw|ދQbսv=i _-n$w%7~ vpCox|Wǽs Mhlܸ`=wR}MzmQS6_ ׯ/^llcLvnaY`g]ޝ~EM|?Cpx]pI \Y]ֳLb;zuu$c5[q^6)J5zmP7JOpPrvaV5di ƭqQlkVk|~=Qo7~5Aw$D[@(7-0w=|=wyw_fsfrH7h}gﻻ4y5h;*m^U\qls _!X݆,,RL.2蜍 mB0V9}20wt;r(Y7!FOakF;.PyNt^*SI?ļ}Xw\*@5wMč .duses|k <NmٔC! dԃqu $KX_݊bfTcB-<.^}M=3T_ūv,p0Xʌ_V|VӟQqykl%DNηoWEuVil`4,X'w gU`5QjSfVS^٘J<0o!&3]zM.z[U>mym/!s,~_s>ІkO*l9y')Srpf훉48sR_& :7wY70Gԛjszc}mu8!;:;sa/QYRYoO)mԒQ7Ղ0'2d)eyÍ"z2;/vZ\ho[҄åe8\ii\-%3}W}VRJʼ}ȸH'%ߦB^s![%咉ͼ~JNɛe~iC:M6/i ;׋rK-h] iuW{Pe$՗@t\6te'CN<_l1zAr~nSW!k8*,RK]nBu&2[ 6oo W5m(%Ro>މf6ޖf>i*q*ۼY3N@fA0{0P)k)Z 4URvo DrFAyv?AE'@ZqVй7ڃq5z4TJn'+|vђO^)3y P@h"aD3X47Fr ;Ǎ4ᣱ&A omTDɨh:s[.5g{f)ǞGBM7覷?r:Cdoܘ|EzLwqmT8;~ YF9: CʡHdb%=S,e*#BsTT懹[Οb 3d͒U%i3tf2 a="^!r'G`8ʪČ@ b !E]K*X%@8#E*(4$i˄aژHTμs1;Oݲsb2B"5g`3|r.1~X݊3x`du_"^S+3Ohԋ0~ÔF<7,_Dd 6 AG DB3y p4y}!rzv$ KOJ@w 0?4D='xLB!ST&(SN1M"E,2R$R0~kpWC"pL>vī3Yc&ueAVm:{_;hM:Y*v,6oft*NՖ++Y`Ӹ2,SޮƼ p܊7UdwּhJ86U[vʒ:Q)M O[ۼ`KR(u‘d9[=A%GI)#O,67o]Pp<2B#0 l1/<Yf .ce@/K`HAe/&d1Yxȫ /5J yC(mHB0RdQTFn-pRv*Ojz@$ͿEp,£h%7F#-ךP:dlVT85yK{7ӡ]WX偬r[%V[;er6.rЦfw<?!%QysYF%0MBR$mahJ;M$;S Y$W)CAU FjOE9p;Jp A42kfd,Uaa1 Ua,G,)Uf+"M`vzf`uO%C^>m|3Gl`B:KEhhD$ &V-29^I䔓K \頄W"jfNcC$J Fbٌ~8 .桠v18VG>2صFj)r1qcLZ'ӣ 3pyY},f^)|=.4#xh'vA$VP'#ByDTH%ЎQގ=]o3+9,yg_4luq~{oȿa goMq/g RFj)PB 貟 9ĀՂX[-EaNeɣ?gu5Wn3J%fYa圤ܦ\#Eȉ PXbPG߂6*b"AdTP+9*v*>}#tL^<=?j6RU?63= ?BW>O 78_vn > ׃?m)gKDϷ_v1$qUg^ӒpT‰Q@-Ĝۉ@ "֔)Q*΅QZ Ë *F *ёˆQ&i#?7_uzFUC}f rp z](gU+΋ e͊kkA7ڃW V/:YP1EUiֽ3흙PJm3yq{IuB~WOpAmj0t'Ny/爁.i>B4DlDD-8mGFZ JЦm24<4a5֕-$ =A0&Suuk"BX]yoǒ*J;v՗c`f<èd4ݷzfxJ(iHQpq=}TU |]j&Nru`g߿³ 5]v1@+ܶl|QǤk:'kW5~}UL_ں.ؒ2"+*]sNV9٫}B>&d`#Q[rMs3ک\Zŵ3*nsṌE* .p\ RE,T@`)d%w8j8Ȯ3xx1WLxwX ҹP]Ŗ~#ei`%~#)렜$&#ZtǬKVp**ɒ:"{C(T`|*+E6Lw(&ɡhr,[*BLa,+M6e:C*Զ{C4ܵLgù$3)jGY96μ?qp&\b>TF).6DM O5Y%ӕØLj9`z&2'A̅A$2q"f|M_.ݮMة]p2?fuy :t{  Z*p)ST  :Yq :ԭʲ;WaSaĞ*x$X̕tu9Ձ֌_&|Ylq/D%95Py0.4g'u`2JuGmj(WSˉшޕwP%4:{0:3ͿfW#C]-VF%,nn/G5.{]A#g'<Ο%)Wa3z/? <j@m7]+V$.:.ZyOi%~4}緟~x_$TqɏDwQ bk1!<9qtwd .6bUNPj#WC 0WyE7 Uo}.]>,+wv8 rNA@ЮJdr1ؾ1.ڵEk{ڜ'+wizq_>0p &J`1$G *:a=ЏnL`)5T!2Z[@ Oһ1Od-8ƵY#s>Nq ci־s81Kߗmfy^=?@vV{UxUX*v 1'Nm&dj:7ua$:x u6<} Մ¼G|!6#zA)'12KCF 9X";y\3q#tA)sytsb:Vl?7Vh& .7mVԏx%k4Xg%{!!X41˜6xGnCU֣qCIT@.DasBn*hT90옃;|#z|\R'_}H.u&Fb^kYjfDXËB G.iZhwMFQdLT,=*}cU6Ǯ8mVqZ'RPl9 h#Ÿ!&[ldB0:R BbB݄R蔰:fD{ zr1W+Pn>{HG0>s+Ibϊ"*?X,>K㷃;c  eDu 0r?؎2w-jr4 P/|cS01a$SٗlT&h8, \Ƕ} e ǔ,J7 :hz–I$3\hulgSKz/d0Tt Ƌ{6{bx57bG @W[eGw>z$ a9& cx0cx="hAYf2Q?$MTsI+NTatU4n]:Ē0FXa f9 NKʛsC!i Uɓ]&R)Ҩu"d4uY!31EgzF&fR C.HQ!7:3$?'G!S}Ie#2`F3/C9j}qF( >-#me޵Q (2Bɘ& $D!h]戚"!yaZ(uJC?n8Γƈ"|;G3\L:2LIB4iL[guIyv\vs9UB}-;/:SFFlrwooa["vC[Rw ;uC~ _L"rm$ME|0.+:C$<+|Ƕ Ǎj'Gp-R=/dZWZg&\`A@cX䫖Yo 2>{z1?{c )Ǥݞ`H#9ʹ XϱChgy0:N[uo)H):&D1gc`!H)mܦ>N a9~KA9o"ðNZik~cӧ95td9tb)J&XN%HC"2A{Bf.b?vAE\/" -o6C/'Nga0Zp!G3K 1zN+bbGȝ n Kb2<`.E)V(<7+RƂ,>CI:&4)I}Bj{򲥟fۨ-)ܾ}D {2'vN*W #V0/Rj 2FO&,QGEm[O7g2|{ѳUI$i2ǣZJח,D?`RKzӉ)pèF^&8^=x\sZ!3iJ) cLm>ycu6 #dIBB 6^O/)1}ͬb ? 6;5}԰^Bn]TI.' e9GTI$dI\hR X̑ |/ƻww :,g#VGpeyHmj<~L'= Wv=0uݽS>nsY mx!GNp5ȸ7 xN;n'T!vY$v&Id EryX"h6`mB!W'?m'~|%ze6u=zx.#*ljǩ2@2[kP)7:񢃁3%]0*b?1^}Aᅠ,3xWR* (/FypA;p zw~X=h?/atDϪ?g؁J UWTVyWXEut2 +_ǃ(K kw”ՠh]oGW)~ v 0aS",_ 2_`ӏ꯿L'$q< d2D ŽwQ1HL451# )CJ@-6pk119{*[!hh6Sq9]9NsK6E:L>.PRi 5) Q*Eve}HG%Ԁ,  2O}D1.8rTEpJ2Usv/x>j "-EvtX IiOnxtR&,:ƭV YsD SЃKk>9l̩$R*-w5g72vU:$X(;BQQ ̌7XfۇavE۱f@^?mÛ~o<;0:2`|y V@ !I( &JEL'"66_Լ( QR07{`Ȣ^lW/\Is͑9ʗxqj=qs y#Vtg͙bJ7 )>r]=!{7\m35h\ӚqU$hĻrtyE^e R\|XWɡ /}\υt{;1|ۄ$sթ,m Kׁ_\M6&_BTh#a dcSF nA28]z$y|txU[e-*L* Sᯈ.҅L EYwݑ8^UIF&H7jץn}<ңכظw {66 zr}7v\# ixU5]jo v12ܯ}tG@hf8V4j!HÅyF<6r`1+YN$\dEo4#T'v*1S3~,/ٓ,VCf3O$HNlxP;0=*\]YpJa38'٨lgmdDKN|t/e]Ze C#roPKa2hĝ=Ȗ{Qz_㿶TdΎM7ӒgeTK'NBRL0B)L8OS߲Aď@q C^hz>jJej6S6Vs}Vpޞ7n9*b鹵m/ډŏ=r D-1Dd@wR0$EpLG# t ̈́RH#4.~gՎcYLC~OS0$@~Κ0Gcq.pA ܒF܆6(ֿb>?W=n=A)qE4U! g0 P\"P-|d`p`^t|pZSo:/\)(m,jHt4q*AAk/ .5 :rLpF!nZi[ 1*ิ98Q ͷZKb9bA=tRTY4Ohɛ\2m!$d4#ưr"bZ; ΢,B: V ^7 #F{ۺ' .b=:촸j(6;l? B=h3d6L1=aKMa sqzώq`I"!׀f.+渗a[66۽|W8YsS%WZy+)&GE %c׉ ԩ̒s9/J(ϋ_o5n2X!\0+' h)GVAh0oP: ޱ搒˗d?}z<7#*p4yGLȔM x`!i(U"Ņ%y(G?w=VQl4gbGFC%H¹'k$ B RZ xdWYĿ8 cNxMdJ*Ơh0TE)5S;_R !f?D % 8!&*jD4ꕽyTx61߮?":f3C:zļS0ZЦmZ57y[ Qγ-djا|g?Ɠ7M9٦"hit8ʃy F- (o=9;8_X"ӊG5%T*x=_J]+ ѮvE/~DjƘ1K oξ;7q ɖ'; _hЊ/'uy{?uO퐄?^9/3i `}{ax|non޳,vfY |/,+YX}T4m5TG !Y[6M:\X#c3(Y'ax3m5m}m{g߶<ᛅ|~J(`&tO;ߗ>'mM&c]oǙM9"Y$l` '&.2I`h#-AP :OED- S[Nz%f~֓#?88:!e",HNE: 4m,7$G+='VD.Iac,C,cVYu0B^Mŷ;2kTv,F6gtU1_!?ֱ=H+M0@#ǐSqȋ*0AEeOOfCF!Sz6vv!HLd̢F|'Ń/_ b LZy O1?WF7?\LB;㵵 !LKrNI9]i_t IP/:fҤObÜ},hA*V'֖ m*+و/:Xf_}0@*JH)ZSHeUP!GBDN=;g{KTGf#䖣|wyvtS̳<`ĝp΃s=Ep*:Õ'>x8_ e_G4 fckw|jNk8O8oC0O>x?=._vj47egz#kML=n2vceoN/bӊx-v} hoŢ+V=q7}e/\4|Z=>G8rNxD3OncWYU/^Gt7hi\˲ Yktp2|6-ZrC oZkZwCxTj%|!˟WiY!.K-M3es*,u@ݒvObm9o_KWb2Ze=W)*oC-Km\G0.'Vt.Qavaʭhan[,{negQ?;~jAMDB!g)ai0`]rYCCN(&o?4~2Ό{sr$`rl EI\QV%`ȌyȓXմc .cPstt,I,ɝHg˒<3`!/>Vt}v99ʧrTuOhyk9axttɸTozw?zպg#iNw) Y?Uɝ6ǯm^^ חjݭm gk q1T(,m]+|O^kY`OmJqsίA# 6|-ױ,:]rIÑ9m`M3rQ(Rk߉yU[91p/%^Z~ݽ-2lDRd`1Ո|ՂIHfA.rkIjho'EocP2"QUಐ%1@gxiKdR{JGw,)%QȭHZ*j))f Ziu@$ج:ۑWiFƮXhcpJ$7ʌ7-nOoVo|4/FE.`:{]+01Y)j DQxd8X'd E4kL=}NIiUaS[-$ T@Gl H1o`)֧xk&桠vq(m{R%hix#cjGBlFPa#"Qttmu1' \jFn.VeUS %$"Q]\l׬:pnԯT**0 "6"kGThcqɶXm4iE" U)Bu6. d@Qř){(\At9%ēkxcDlVx}ZU'J:]q14Eb^ "cHX;%Ԩ؈֒&!SFh=.Cð¶.rk7q #@nNp+y?`[ǭr|jۿS'B:/D%bgP.Uq=sB@K1/ƏᩆDօBD-!*+PE墒mC@~\iPIc| ~W7-~/No%}\$z:C.bvI{i&E]W"KERk3yH],[3,^夿h0uŮ+o{V;I$wd6J瀑Zܙ0g lW)1%uJSwӭm^4M;P4%HXJ2SF1,p$eaO'CT3؋i`.kdfܵϼ`rf`0хB)(mRhS2$c.@D!A¬֛mSj/5 lރrڏ%vۮdZU{Q93 ɱf}hE.ِ;2ś{7Q{@MT˳+&`tkRJl0ِVR{#&~5Ŵe (} >"JC&X 7Θr\76Fys[lgK\~lvv~k~a5o糲 3}(>|ȾM`'/$端Ws|GIe_T5 -3Oa tZ-=UjOU!Q:d dfs׀[]B65:2#?2#tw yJeV s=/3gt}<䣧D1Z-\d]<ӛ`E/'tX` lr1$432ԑ*f]p콊KYOBGX*#G On y yL1j B2Jҩ8P[N!5f-âbk̼w;f7*~p))P`"@'%s6ϑ۔_pJ~&<&G40;F\ab}HiŝGV|4V8c8>OTsƷB&`o̸͂iPOlgf4s;o||Unj}g 9bFM/)s+pUP8Ml >?{Wȑ__3VkYc^;n4"/!?־>F%H*JTPKdeeEfED~GLaT?yT[}+y L]2NS3ڢ͟klw`,Z|6(6>p\?g=ˤq]\uq]\uqnoI=roiu=?R=RqFA]uoA}Ph}PUa>碤YSYb6tS-F/H3WtU)r+cr̮Q=LaTTJ+fdgsh.^bIf=B;!`TRNa)ţVT.} ==L2GD 1L팜=k6> ]{|5筍UYpYﺻ`Dn%mVyw]z-D\0^MQQZ/Etg;:]n~lRˊZ. }mݝ7.{r;?L'--n%oǼOy7ϦGLW I{oԵ5_6ZH硺y+F-6!\ZklYfo4گS л+gL[!M`!RDzՏO,J{U s>//tN`ՆZZ٫+92]qKMN2ɨI+BoH>E2]@͚䕳sMnMm7a^ȄQ܊j5zzv"2;0e*sW\h 8VE̸[L}Ιa=.ce{ߘ{9 ֺHAv/щ.`\fWNWcJYk1ӻ|>62כq@l ;lg>9ȽѨc7ԣ'Y_l{BH˒u&)'ӳ_Tz%Gew]MT Ʋ`bE LUH.zwС'>ydG0VYry( qh98,l "ʪNs^} 3Ic+X,[F<bkFvIZd۽qʝhv밨uqˠ#Ì%ߧS>z0;F>X~,=GȾ!Vd0[ 2|> y>#dDEXR%A xʺS\|&19tJX1M|n!Ʀ%hjmAMl0Eک$iCdAV&hOȌ1Eg<`5Q4Bqnb)ƵY묢.<[Cg@b+ VB= $;U @:&eyqqv`.C(S Pyn}I//~lAN&WogdmThDI7szpt>fZpa˓\l-K?ׯpۻH`-8I1M0zaIS%w}VF/.Wx|J n0O5n&I$|3^}]~`ӝ{wowU_{Z0XCvSOfi8pPr"ioZ; <^-;ɯaoIh>87Y$(煉bԄ-Gxcy6@=mf7=eV6mLF֙TÙěeqXԳI\.="kl#z7"B8\.y^Ä&qph"YdXXAZKIjh"NxT)$ .I5Zir!犚/VBWw0.^e8^EV,V7-|Ysebfd_fr]mtuǥ<%\Oyixdbc`-WI=Jܭ:6q?mz,:YZBneu,{Ѽڞo!M~Mf2/ζ/\֩hȽB``Zd &H@ygddf,]əĜG<ڑE^`s9Ky2NheLkgDR$] IEI-]NXkޑW80j7ק38$]齸'+}bqW\ Tq}DOBI3 RTl2JKûN(\ ǕU PE4TN`*q*P$g>dWõ!4l Rhm2Iz!& J:g̏ ٧oݘ^!}[ۣפg.uLe#Bf^ݹ~*ci[O Czp;zW.IR=$oJN]. i.^CdH yL^b6ӧfL˘NT衴tALcd2ș@>yьMg;J$nI/bigeM3t7nYuig|gF /t GZym^{H"8Mf̲<}o#¡qtrkco@.DasBn*h$T90X;#goYO'ynpjGȠŦ -vEoYeuI"B G4/iZh.;&Q(Y2_Z&*Oʃɧ*Fbyb$:2/}&MK+̹ \},Ŏ;=CDYX"CXrFA>th̥0< C;m)sW$W5G;N?=H-Lu%yf:gPpL zKv>'ltLB9Å8 \GvFΞ{+>?{_XSP2 -44[^^F/>xWZ/vw hy-wzZ|W]֌8E4-+E 2AUN6!sIkN_Tat3+B]%˻,u d%auYazٴd9k87$ I&$mJqǚq/ҸPT:'!{ə3a$LJdȢ!8ҢoT N 2DˏEؿr'A у :G!Bx 2ts1 å u+HO 2F%bK sdDф<+Q)sgg6 #[3i9 z$ #Jd x~^sknMt~QPX-\C`qx[LEMԻUjz/P}g/ՙ.?=i KҎEq>L iPdG 0+#:D4 .N2R/>/Dz}p6a.ڷ17Ry o^dL|eJe0o {s,pzg\媅*Xo^>O;8VIݱąQ|;!Qƣ1o9_'ǣ[]kor+?"/Ff {ܻȇ~+%@{g^hi(Q,ֶiLWW]}E+o\0˻no?o~]d(7׿}mY9_sZ{gaW]LI_4R;RfOonM?xGg[֝]/vi[`Y{~2q[O{%y2"ӯ鴍c[}a^Z {щ]/v5$WuYKA[/OT%__~nK[KlVՏ\ ]Lb/nҤpqfVG[6- -~Ocƾp ~Cg ugˍ=[WIċN2t6~\流cgpaEe)SƬ'_2bbr f4E͜]BH{dyɦ>ɑ+D-h%䋑I'/ ,b@\s.(Kqj R&SǎPygS߽Uw7[#=tǺWXOUŔ/i0waO8m`3?<@P4:(QbC9ʚ*oꂘѐI!u^9,LtslA& }pGJ3o(!+O.J] iH2d5aX0D>zJ9aF0~#sp$5e1h]Tv`8XsD;]ф_O8uPl`Ce5z6RvtFF.BV  02~ʌZ9@ex6J#Bb TddIJ26yfe'Uޙ,U]L{Ѡ^yPѠd 1 EE=$$ P<5[%ZK2> w()FQ%=Tzx%oç2]VZRОq*2 ЌY|/*2"]lUZ~Y^hp:,Ͽq]ʧ*0mƈ`[r( w쁈aG2D(։P7P{2k~ 1:FUuںd!ZN숹&R!dP͹c,_cv#6obb%TY62-f( sPdDnXH:L̐J`tqkrɪ| K3,phis?lI*cPuzD7GtG=( xx[-gd>d(c昞TTuP$$ Y̍-Y'uvcPSb&j=`͹#1_\rZ`Vr_~_ܹ]kxFNyu?,M &jBX<A>D<yl9^hrL~ l!Yu!DHT2@xbV*9}bq;fkI=G\1.2{:At\b@E66F,2wۣo$P4-aqwt}^fy%W^N) S[{>kRcԟqؽJathTn@x[bAbj#X ,4$,EG!#WO%I&*?E >Q*c1V$yvz&9dq,u:Y0}<XtP *0A11k1QHDsRZr5uzT-dľ:j7'uHsdžUx]vOj_ Al.O`|YN8:$ϥ{ Lp2+&YٺSO_8c>]' ˙ lМ/Φ2⑞M{93GtM`lmB>y0Pv=H4ZTZĜ=h fQU لlPr#&1)0|}kZKCZzy6K+ro^ ,i-"C'頢1H*Ek}BkoP+| 64u_?iOoi{M>v9ӿҭhdxүsR=A_Uf?.̏bm-7nj[< &ߗ\vaTlɐ 'L od…vLEP SJ)dGU5^v-` Of ǖ mˎ1eW̋=X̋AΣ{൹}ƃ9O}n{1^,vnb29z!nG^ջڹ:KokɋAjȈȼI@>PDf?Bg$&9(>nU6qHnCw_Lsc|ɧe69ru3B.-@x9\6'1>oo~f[GmmY~wec>u%zU#H1.VZ}]m]q8D{hvDZOyֺTM Kj&/gs#=5F"d"H2jwIY*$$hC'8&zlf墨z8M#w6m9OhΰUZO%7$ۥa.pk+z~>)9=f[!lJ{jw5o ]yG͗J|ݛKgJ uexQqU}M 0'=}iUt-GSṣ :}lmO{ ωaǷ{gnGnAJR]v0ODarGlf|SA]7 s HcjMXYg 2cH`d!i;My &}Q\uI [ҌeݧzrXv1ÏPҸW1}5|aūk`T !,FbD'b56F7FfGif1 ΁(SDJ(#ےe20Γ rx3,"]lq{߈>l ]]N]tއ_\qF+8,:<R%BPT ";‹O?w&WaEDRl RPbрU'@gIy*(S$%sg} sڦvlkvN>m#|"f4`kJyrv^в)h`:3`lsnElx.S ;2F6X9c$wJ2NfR$d1@G>OՂA AIJ)t% mKJ$!Hr MC9w^蘰H;Yeq`G,gHP|W}8;jK\/x*cd 9&@䣋2*W=/R O.s Il?XgPWpve͉XQM W7/9-GDA2M%q_;;9Ey_fK{4J>X1Kpર# ]҉@o  O?q4'1Q u9P}F0RD:4Tl]`H'krR4a}3("#T2%erST:)n׿V지 -JdFnEKn?]pfٻ6,+yb_3v0]~p B2KbF}@YL.Zh97FܥΊ |51t 8u |Aws9T#mtz] )Sxϓ=*yng-'u@ď_&g\]_*D)}ϐV>5wordEwv u cU AN3p0C)<)d,"Wvs27(Y>y1Kr!""^P̽ScG90bNg-Bg+6U*4,ۘطap~1HOtǺWX/cEi %W^׵+9?ɤၱXjT ʔV0Y߱ P[<_2Oߺ8Gx-!|GcG8q@/a}SGߓo4 UguE%uOpwN(xEJh9řɁxePjlt<3>z:G׆<@A TFim;X)*IE05 S!NYqiד:蘳9C"Cۄz)ʭ v`j/4BjøQKd5LƀHSQl=CD![aP`sr΅'cճhĻOkeI`eQfESlV&MJ̎_tӯsGCa.ӯ¥c\C_AY)/4<0WφAv>MMo&|O0;.A;q|e0eڅ\hJxkF&E>};.6 G?7&9C5&w?~:;LOφoߢlTurr?\z ~'֞hزZ0NVCvQкp.P0PhuӶfG wAAϫ.*>IZEE 6V}OZK4ƟֲwUhLu@?%9ؒNNT-nE hbMo'O{t1٠O/n[M9O*Zz4~Oì5q]ͣJuyBdյ{z9v2|8R,9 =ɕɝ^:݉o[ =̅?!sŭX`TjGִufmAy5aSSS}i23HFޓ-%M)tϗhwԖ) 2ALV]_6HԪuObj3y.ƜbI3±Rۇ7L}< MiAiyG$}d6hEо~ăN훵 $D~G;Bm-V8<=ܼ'_, ^Qn<bZ$H3!t5Y-G~g'}gHY"e:`R ~GzAj LH} XhAvy&1qR5&%3UθT2iW_D|D/O ĩ7}{?~!lnˋR2l!s%#u+8eė q^*l0, ɨFye RR&("4_AH2ø%Tr /H̼NdG')G$=sX}h8nja R"#m=\)*1Fbd;B r2mXWxeOoh=}}שn[ztW%u.6JɌߋt8߫ϠbmdCH[vĘ^aVhVDϢ>R kr8jL));%'LhOTG,>j i AHRVa SƂLuz-#ckj4BZ"2slSa{C{Qa3TϦgƩUTxa^b)VOb:P Etp 0* ͐u_ Ѐ@@9N`.0 DʩR iHqs}ih)v86+Wršm< +i=N Kp^tŬ r/|6.+ŨcZIñ4N: a.E)M= Tx0/D"i'o۲WO֦ 0sX/;o ;;Yll෽~2v7q|Bɹ`1X9^b=cgmI %ȗZrV:AfxYk\~{oZu{˛"X«MFv)yz.O;0xN`NbzB|(QvUCR&#hl:@Um`H)}ɘ!▗D {S՞)U5Nyʓ(c邹GD g@WF9ob^JScRmGs+:2,"1r8()@88۩* ty _ɁqbVVegreE7 M,iI1e-JVDט ƈ*R="qv~7oo_}%!/G~G2=L4=Kq @+ϐz;L!5M`*_yyv@s`W1`Ҋw>C^ZDRe-9Erу?>?ZH5ǃ~8>EgA×C\zӿ5Zb]Zəe A64՛Whz8(Z_/ Q07&OW?~>Eû>ZT^DΏ}qw׭-2W@ _Vۿ|x !g) g?ZH{ًŗBu#__7g<[y;9moec&:g'UN{gW xo<˭uߢWr~*w_ZГ[D؏\_Wzf]ݔ;j{ ?5^Naˑ︪_7m ؤ=ڂA?.|zw6l ؔ/gZn8,JF?RN6R{>/Pߊ)vg|uXcU J=[=<=+hѹɷOC =EЭWs~qwX]5zvvE>Go?KswGhM>zWhW2PFzmqkcwƬj^~8yp>a΁@߭=~^PwzUv"Z s͏gw 4a`W!މ97k#ۆyOGAżp潅^|[/߾Eջ|.ه<XIO:~Z-;s7>#v 2d͖;G|w[q:'=m6<¢}I/A30{OZ]4aZ+R (DiXQ^Yն xLt%Mw3uvgwpNOMntsGLj2}/ #1_YۏhvZ3;׌߮{W$ %֦L׆2h]o~$߸6$ã-*R;W^j ƾc' "AUjot]Ww[WIvJO-3/2![Òlx22 ;\6V"C:l`nk+SQ6"tPrX,K9pw[}'+߯~ażFT :֢k5G]zkRo?v^1d ٛ㋿sp^f֮7_y^ǵj^>Jڼ*?V]fMݸa09z:Hi'p,2~ЏjkI[o2UX9yϯNkz󽻏;/A7&4Aճg}z K/?M:fjVu猪Դ]tFs;>' MAn^| xw}_apdyqyv}t:XKnaއyV$gk>~oWk_w#nL3Dz㠜i(ۧv8bS,HūH:{]0vt{>OG(E?Xycnv{ed>AA_|:v֟_5N'v$ۮ]ʷ(KFEVƷ*%YT6Z*2[Ӭf5%V=oOk^#0e١/k[Jė[\=/:;_㕹jQf!l%ɉѪ8-%h&(AU[\P)&jQBr.1k-bmŔED1_!vҏ9[1H#ngrʓZWkPkAt*JaԭR :KʎZ ׂM&r_5K\5#7j,N@R: JU$RK#cri Lf)e%l|SVDG2eQ}h9(+I0aOFK$3R{}uZ-#C!uHJ~t-ʆiNFhR٤ CdJ BSgO"ИUFҚw-Yٚ,~%|m==W,D!v0aʈϩY'1כCFc-[PCB'Tɐ>% V?W'erZQdUY:QS#J̱6gD$GYmtY"ׂLQ 30aJ?g=A`ke#8:BӏR:" -8fm43¦`$ E$dHTeʦJF|VYb9kOeI&l.-6JR`ƨQ1K"eXY)dSh! Q cJEw)I6 SI*e«@4.ZeϢ rNc%'D[Q@Q lzZCۆ5`;8FQ͚H"Z PrdKw%̣`,.OZF%"14f,&YgqP iE# Jʞ˺ FBPFơ)͛ݰ0 PȀop{X EW F#آyX|.bnNc[-tE*d$TX$A ^K\͠ ߅7Z x'0sٷX="z]C Hd!f@57._s*P+1h aƣ9b΢Y'܄#|( DU9&:3XŸe ` I&- p@P w&#%Z92zܨ`5+g, hl N+ yJ )mveZs.5udED_\lZBӘyDV*B¿(Z%Ẓ q@X. ȁv'ІUP:o3X+~MH8d3u&I|/nW4' cDC;L'@MDT߻ 3"wq&~'NٍV +.x}%cc;[x*As^4P0|8FjQQ@,W%ׁіFA˥{wS@yE nX@hVcAZ=%E@i4A2 Yՙ|RT+w0n,9h:{*ȋE%YZ{[x3 nWz8e,XNU~t}%*y*C@Vɒ3LH)qcU0OGޞw}2W)Kb,חXk#p`Whc٣.)1H/f9 hT`%P\H!Q#TK^HS@ʀv0顔KpQ }Y.z r l+ hwX3+f4 bduCJp+Ӄ"e#KHqص7F!uf-3%Yds ` BAo,Y 8\-Ja,0l" DŽtU"&B@:Cє Z(w)dAw+ 鬞;k4Q8FJ4؁l&[J&MVDEK,­nil?,f$l@0U}o^%0 R FhӸqN^yM/`t8^WleYT$ ԍiFh%FBƔ(°4ENmQǨM[cVYkA#yhY-:`4xDp31if0I/ǣaʷfqs눱 N:)r]gt*TE2H*.mP􊀔wkڬaz5٠*OkV (4+TC(MxrUpՎ9-E^2_g,=ԟP(iQ0RSd: f9Z:p@ أtF$ !AT80v*`c37 Gj ǚ*Ei H>{jfR&C` ZTY[B):9Y%kP|P}b P[;ixCX0m@xxX@¬aFY&]@-L0E zQBbx~)Q*`H g ^S{(0q˾ % Dʭ_9^#0@\0 ƍuF6kS֨UѰ\C7 sI8Y 41o#I!:.P,=wSJ-0 ]Z#AM,?"u-+zG[V}TQnիo괝]Mgg[QNtuI22Ko:H[[.~9}KYg"Rx9jj`|~g{O=)zUs)d+Hm[i+>E%b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V}J |@@^*@$+>E%vB+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@@}{>J kֺޱSTU`%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J VU 2zNJ c|FJ kųQZ瞼XM@:i +X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@;ijG=lۛ?~εվE>pzycZx8_عd~?xpf@v׈  Sr9l'Ϻ?pg/CRbl %{k{bj 8ol"R%SijhIn<_NZ2q¶rr{oը' ^Wa hjL7 #A]qJs[9$AN>D#ur5}1-" ^ _tpj@kwls݋ddk w*F}i}3@1~]Az9OI˦ŸXhu몭5EV&|ۅ0L^Ѹ%6& POV, F*܃TnW~؋-Yy'&'4/M S- Z 'ָdq}EUYs.BU y^\vZ=:+ /4ſD_ J_2E;Udpq͡ 6Bk]9y۾628˕` |[Ό9 e*"Y[p\WQM{3`LK(ՁE SZBeֲ@F^FDK`:.+;qgPby/UC|  D0^tUZx3McМVrґqUVkb/f;u Ֆ^SGՄP;h\OB27C0ٰh^rÜ9Q4g2[U6A[鐿xϨa9ccjd0LT=`s(@L>圵{A:@-%NZdmP82&%R E,u[Gđ1veŬ[+x.]к"4Lc5mqpT 4iɇ+R ـM fd0AT'Ϋ/[VJS|5lF;(N,,f !m Yg/SY^de55h<ϳrlxu,o?7%>8x+GKWf 54}4W0`6|ߞnoNB//ظpd)FjCݱGEw%9p&,{sl^]D-3V /S,e{_laFF]V {U+L0RPk e?&Q_LlU٠ԩ2btQ}{߈(m4_ |'[ gYNͫmlpEv6ζTh/f);T.xAExϔeÏr U/ uޔ%)ĝ Vo$!"ɂpWNqq(\g q+ ~.|,9,8?ՔQ⤒#gEJF%S^DI68-KtSg8 7,!_߭nI˴hfI}5B)36~iN$)+-PX ǠL~9 ZhEQ*u)4K!+Ed4ۏ=.'Yg_Xy)S|J,HfA\BZ)/kܑeŁ {N1鞩(B:HI윰Xs 6F)b̳tc(YdZ3I"&Wbzlh9HIB_ M3~5u~?+K䤘drվ/͓cѺJEmO*A4[=ΝIė~9 {CCȟ@ ^V.b헹Vx)#/ 7!š\;@"Y|f6fs_},d>aPN_T4,*/Qlc^6cwv'8A{y.dӫ_W/]ZD;z2&/"}|ۭ|-6QWkۦhrvY6(Y|*O>qaXJJzwϗz>(Oӛ ,d11+ۨM`E_rd ?_9tݞ n;E&~!p4[tgҖmt#ö lmu~[),[XϲlmOZa%^e s)UuڃQjQ@̊ llweshX6J- -;-sR,{%wŦn%F[]^PnՂy^wXw(aUؿ\w i)GSxY;\֥}6+|`}^Qo X'uo8SZt#@Om9+'$I< n,"";8v) +{/EȁCʉ񀪓2Lg.Q>hٚX3# J$᭓04o#;=Z)kW} sB糺EΕur/,*S>tsmCP-ɡ8Q/!1}ŬТg]vu'PԵf7 i@dGx !t (ˑ{7JyLAM11h:RM,6]b"0B(gJq΅'(.<=#q`ZY{'X΂aV:]Ȥvt=rC!D]1ϗaoߧ_KFffɠE[INSBOt0g&Zzz1J#-?߷)ٍ4N=ו߲ДdÑ m,u >bgwrP:&B]rm9pox?U@Fw:o>ub-l6=9z_,ƻyDl10,WnH!Y},u\hU9kgTd4mNWeT$ddwE< V|NZ6FMݷzՖT=[GGr=(}{"idۗ>cϲ?yIW@ap~׸j^~`.Q D:E&mx^q]EAL(gI dU{ f9v2)\Ay(bķ=Bk—_ܪQFeT7=6ѕom6zc׭SS^0 VDzhA|0'5hߛIjbp#{rF]ݟPԨG֡ZQhs|'Fl㛊L`y`e5v/oFw4u~۴6j"CqKGsAaĝNB" "t_P3r3ʹ}+v;(ͽxlu!6$RƙV: bRѲӉ<0O$y"utׁK x; Rsg]`BGB@ b灉Z6⮉RwJyxe" jcYvR,eI*Kx]U`}פEA/a" ! 14[Qe$X&%uO`_ }-9놄: lrՖP|Zt$6i)663m,.D{ן ;NMi#'xt\KxDr҂gL1+-d Zg<VښL()V"Z̒ 2=j*C)7 5sJTk[jmajdʶPItEAvmpw4)o Ǘó[ewAA^*Ju03Xkezge s]dAe/ZlieBJL (fSj5 hCɷceA 1elugbIDZXm[nͪIr*.(UM΂ dhȫMfXNPGd8r=JCN. yQ&CZ$a&JInf,Sa춇SH-X,b5E4--bou[k> IE4ZPN|4CiI,JX`C"ērƴ]ֆK4 r3Cdc̥!K[eX;-bK1:I]lmo{uVъHSwrV*3Mp \hqM(QKZc$!jq,8ޝoȝ' ϷtimIp}e?~#Q[ga8'j)iqvuݛv}͆zZ:?ڙ_$)㋫N0;hI =nGŤdBztNz9zW%o|2e2۾G5~u͟rHS:Xm&H͙0T}ƾcL& Q:f6&$c>$BN BW9xM!p}+:,.f3XŇCqdgWh"<2X`IdIj4tP'ZtN+o坃gRz~>tچ,[6K_C߂o{Ǒc^9Qj  6 >WΔtM`J5V`fֽVNR,3ɑ #+맸* xPA(-Hcp@W %-I~YՓ XTPq)=AD*@ JY+.b0 3z7:5ήxWU}|}w}ߕ,sYILRH B3sKBE z8 U> Nی%0@חx^ZydVc.1ypbp5(1BVS0 lIwCΞI_>^V֖k[R=@TT##{z6M^s}3K[Wϖ>zz7 5AAerGM}ԿiԟMrMJgDNNE*e桔RMne/r[[ pr]6B\"B_bXx` M՝ù]~Z뗷c)&Q*OU{O >@l:p:#UiR܍ P5=.#0bE WdXU=TjUhBFE .2`#.] er^ࣗhr,D.TJq!vt>!Jsΐ5E+вhz];*Y暅~&)*Ym=mʁƙ|y$+py| bs^oC@W.4 R2w%B*mqAJpf\-yeҝ0yĠ;()v$}!Ԛ3~7 'FT)jbIZ.iB^g rk2>gKr;--&!Kz&'dJ$ݱ/_rƚCT ilF0& y9r(x&j8Zw2c2YZÜ3ߞ5>}aZ/h"O9& >tfr%8Be&2lP9Ch):},xc٦Ս'+>r1 0FIr5҈Ihs(:Li+>Cr:S[34D p9rל![ ndЋtlA2pEeه B:o;G!bFmBH /\Q.C"݅)<^J(K1MQK=Q{M](h)D(W1^$ỳQO>v6V=#6JhxF4%chO6O3|&_ޗ-خ''.Ȑ=Z7Jo_[b?σSWޕMMh}kkw{iPt 3c1 fp۠ȏX^cf4Ps4bHND0e1 ƿ{eUF?Mgo>,K67nn4|4)/z[ H6ct.\Eb ɪ@3sB /^w}HW;$Vu917:MNFӘ$ Gi Q?~~ȑO.=^ϋB0va)o_~O^tz&o~FfzvJkscFy1X/;|S4.v7MR~cq.‡P}Rv'lkvsJp0σdVm 6o{gͲ|öY֊+`Y "TJz]֠[_^ ƣo''oֺ$6UZ֖Ztcfʾ\-z3_<$gmC[>  -v0Gf%o>3@X!ؚtKayxm_oʋn&p[vVxI gf( mB+qYx Ӟ_h$gCNFlZ'GNwmHWayxQ;si`vg0ۍ( ֔,%UՍ |II)[ʖL8''N‡T,srحfCvN=;zv|xE6yFq2N)ڠ Eq~:9'I$ǂed8T6}oz0/ܜG4!qc7nc4f8/.o$]樅U)ΫϧO\DofBNˊެ ?FX0;P+Aydf8,ȼ~l0DcS =0<0J T.GL>$ >(UI91RQ`Z$1lYj!H4NΔt SDg,BL]`XsXB !Q@s~3dl6Wkجݢ,B/5>:ߝxRÈ1Ε%|,prfQK#`Q&rOSx&3衧zzymȃuꅥR45V&?gprE 9(Qω쑧d(;U{p0r|Vor3DD+e5!Dfo>z&1#'h;.zG!}cEn?JL65Qg7nھR 8%+qg h"5B9!ij&; w%R&8jLJA㏇]Ջ=ۅLDv#ƳNxU[O%Mc;:G?!qh~o9ד;\i re9Fލ\>ΊyumswrI~Y!o0S'0?M꡿|d.N@ DG4~XePb.M'W٢EgOkfol @糮sRTN2v7 oUo.jmP;&zwS;: u{$EhN͚9|iwOܐF3߯vlcoځh8 L~<(P<ǪQhSUw~9Z-sss eDg6ǕhO(rwrkR')U̪ŝDlUVG,Z1ЈYP{\,3]2G`HA)(Izs ƒǪCէ٢,G)'d_O3_("ivm 1OѩCJ6΁Kzn9J8.A1]TBKkbY_j QAgiނ69d=FXBtR.+!VL=\?+*-Ay?⸺_c'BTøڭs=nNJ?aǪkg$ ؝='=C`cA^hf>^:^Q.9rH0P U-f1=8;w -u}HxEک #VrH̵i-Qpyt.*Tqp;P*ؖr41NkAbbĚDrܥ>¨4(sV+B%5hz9OƬ 6.bQU||n +h_vx>Nf*(asEg^g,zὈ,X:bH^?bE(jRΪh)U:"D+sa9|Dd1՚1whT)W؂5g?ȳ^O'Udy쳵oA~[>Eh+z ?< 煨b)S!k;h-#R "&M|S0ojtʒ3HYr޵-݌%Y7Aމ!k:W1&'6x MBhM60C>4 ,% trOEU@'2\g:PN[-:@y1$"i%khS@BE@YB2&E>qu*:*g2).YwHbyu#+1O U7pv~RY߾ͪOn"EOu[Ӽjlr᧫\Oٯ{S_}{{N=uS3m识#3;fɚכ5nvtQ 28 Ɂ}.ё8đLZbCD#D(tΤ$>pAJ vtok7 5JM?Opxh#2?z?!No*x}۾[-:v|RyN5H D p8 nO>H#@!(a S CxK)ޭuO?j,y0 D0B<ȥ1ф'U#bJ$Rp›Pn!km]lGDHj ja:炚0sɲ.Iy+7argjFz;$kt2)̄*C9It@H<a8E*{; W~[rcn;7VC\o$F ʲ9W#<|4U;йQe6OКէ?FP`JnNiQՉX/9ׇQ}I9Jʡ%)gǶ0K@:LC~(!PISeQ$H%5'^VyHbT21C7^P!8"RՉa-wVz=SJ5gK&li;%g92$¼ O.R:-):=;h:! !1g5(y*OZ@ȒD Qx4͸51kfC&i)g?xj]1bdW0y ra| dKև☼l$pyɍJ͕)Y;;)ljWD-jYF֥Hdf&!I)60+w ;S Y$Q)c qzԂgbNGPTqG.!HFbٌJ1,,62vBU e{/i̸Cor~!ۼ94 ;~BaW9D& w\(8G0F#% %Qib2/eӚ O!;{=,&W:(,脾161Mªb"AiMd +.桠vq(zښ>r I)UʻQiD<E &D"J ڤ4Kg<`DRͣ =i&4!OFbٌMQ[gm!CZ,6JvE[M=.ݮ8[`%򝔤LD!"ڔV;C'*H%j& >pP8< lCuyfܸYlRrM\9ns\BяMAC9zzO@dL8DkCEtZQ´U$AOђϥ%'[KN>:?M 2A%*m'9Gݥ֜B6"h҃9e+qHgG%@`l/`b0>d-enuIvu7[.%VEfe_ 0qriՐ;چ.íy]Pig\d-HDvG\X`IJ8!J:#钍ap^w޾Kio<_4<\rnb 7pLvkՕ,N!1Ӛ(6ʝ۳r+gJbJ5V$" ݗK'}It DPF+!S{yf< QHX|%W)/N/ ݗ:ٸtQ+ i݂Ư`uw^߯gt|i`vq=ADh& % VbK,LCKb]%'V1-6EY>}L OuN;6s6a/orח] S8,8&mu}3C[G-[17(7AArrHQuHQO4Eed3Ӭ@HLM!gDɁNEU*eʪ_v.q[zx!q]6zLTH|%HStC6I-av寀.pc|Їjz>@}-c 8KUrdsg/2ImuKsY6r5Syj=U4/G3Wc-?hۘݻKsi.ovv\XlYv ;\Xy Ey|MVAz0^P:PT0_R/?Y{r"dd\a24nO 2:xԽUk`Yze6t!]'tmہiR:zNcә֦:e} k*n ejjE':5{'7 G ^i=%3Nj o[~A"4OR 8ȣRI톖b}*j]”[&}f友0QxGWo-NѦɶos+3;} 1o,u7ݏt^Nx[{6w.k]œ7?HxƝaruy˫ߵ]y߫ ˖{.Ww<;{e/oݷvaYlyޗ/h2\"1m_(7B'OC^8\C>m4MAb$RD6厕;Kg8Hr'F1r$M%fi)<,N#qN.JAv=e:DPp3%I !P {^c^VgBS*\Q F*UjW;{5)Ef$Y>\#" ܤZPȑdV,!Ty\z<:D, g Z#;`L` E*&k&bHMCp3|g9%gaRp*a *eD211\Y R{)erdlpp䡵e!<<}n?Ͼ;51/`ٗRa9] ~:' L!!mhlп o ~n1L+:':=\3}w, s}.ў sIu9NGo>\2W)pv!'5]eţe3 5:1F-mD%rt?>r`/O<}s?bv]?K<ցp8Cg5!{MbYz!ڰnk 6n0]|F<[\5h{:[(3\N?0Ï{7_ Gޥw&ߎ o^Z{Y'E[n #˹gN$6.Ozs`~r0V{$ZhaKJha&Kƞa9e͈N?6`g!eHn@Jƻxka> #%'tE  0)hT$.Re~tj&aq0I޴}f^227'&b/ q3ɏgߟ~q .qRʍ~y6[f )#G_Qoch-Ϊ߼(x>>h bzov ŭ5-s>ɫifY$ ׾RIiE*VR޳Yv|~^G $ڧI縷JY n66N]߬3qL' nϛjyh&6ա;v b4x]%kdz.e+ca&wrInۯ[Xm{i|4#Lq-㞨zDmN<9%d-َGLKB-rz/|eRMǘ?up Bْ2yœ\GZyGȽ$\m8V;#t;k`Bݑv?o'9} ?-`Pդ_R\Z#MUtrm ray%E2Vj :(ĕxR*-6u R>,YrpL!QJ2L&PƇDLRZ9q*%㱿kT?Ǚ7|:+-J$j˖JQ* 6E! e(*CddRِSό0- s)aK3CEaEE072DQJ `t"֕g$g0fr$:̓AKc<$n0v PHSwb$!aM":ڼIa<!? ֆhx>bMOFcZE fsa-8 +G3rё}'gyv}c2ET#-QsvV6̲!:KX<1&Yk?nvWEnXH>\R82BIsŤ/ eU0A*Ҏ} K& 0<8]rgp 1kO-8p;-s7JcjL<:`yUNu*c HLy<qL *L:CpĖqaj ސ뮺+v5$J|3pYf!0#aFZL~mV#TB`uk>ED2 z'yF%= _KX<S؂ynAVbQI2a (Bd4!Fڝhŝ]^a Vev LIU!6!9`kmPeSgUĐ){},``*+uT,P;XD@0Rh jn°pfcu.1[lj`4 ϴ(I]l[HdbV!8Ip1z@T$ 8G"Lp?~+ )dl%Z3-A{utԞEwi<3`J f}޼hPR*\ץ2^=AYv hߨt7&& :K/kJ^J0Յ!r5hX 3+ F~w=va̦ysLI"k..nh!}$0g2fg6:U < 7mbHxxjѭ>Dr9RV <#rt߆v|Tu:A x ]W9NP.#4nkh#Grg.#2PP<*@% AO p+r=c5X߸Ya]q5B|V$q2ΥxrpIƷ:kM'V0 .luBJ"8R4xT%#.b: ;PuLc-й `BDLJ}.S4#70H 5j*pT|d6+0bLjeNrqlIEX'dX_tB%K;H* Xk0m@ncqX',0² B3 _XE]biDF: |'K8>*HmFY|*5LZWM/"c`aVMVQOM`$1Dݡ Kq)z,Xc"Ε{nꊄީd;\g;#̢ԫB=-u-n7ķU6([+HN{ӓ~>eQӷu ^x9U|a Cmv=⎇ip`wǨ*T*zJJ ǣWN=%+`S@HƐ@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H *G}JJ Q\m@rR}J΀=@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H *LOI dgOH syd@VG2uN@\sR@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)>_%OI dQ\+NG2MJQ p)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@zic7yô֧fqbx'j_&I|(8d>_<;|,tvڮk5s.CRL낼u.vR9!dDKj~SVwJkx`⛿_-49ulxA?Rg_ѳ~.jYݼW}¸ܡ'}AH ZXg{Aư&A5c#׺]҅`2!|Qkogų;t 7s~^LU5Wh1n ,eWSľI9iY̭6uOgV|{9bV mx -" $ZgjBc74<?\r.S4KN7kl\\`wtqE+gGZx[κn.C})Ol2=_*i_ڵ~ny?+XW4vf(_pEkE*"oHрfN&S|֟w{nY&[sjh J8Dyڜ>,ՄA\6:ys)ō~n~. GJ֎x6kGreY?R%C_K?9_<].,\vv|q^vNbߐT׬+Ǔ}'Uw-HWz9?X `t?yݼvU躈15|oS4#&ލ{:hwނ6jxr|-pKg[J|[[Tnt8ty P7ǯjr;:wSc٧`= =-w`k'M,S63ΊĴKJEW%bs:"VNjS*D7zV|.E\6]U{?Ugkxohpu~rPsa\^r-Ur75/ i{sge-{wvb|1oȬNpp]W,$pFD1n%v٫fc7A «Z;-GǫiA*츖9;g?TEfI!RAբkd'JP#>֪E]M&]d`G.Z2(* >z uo̪|~fPB͊ojo3/%g­* @(7:X@@g.MB?v2"]km3$Uv(G j2X#ewe7CBr'fR)m*1K!)KѥS*ǹr/u â'ˈ8q4snƉWcr'QOU-e u1u|\^n)?1̀no^Z5rԓ\ ~PEUEfu' >d:]xE=h|pۿF,i0)* ڈdbcHuCTZ;69,=rwz~o5ha2uGg˾Ĕ  uEv/Zf)Ks(!mozQ_-kApO;j&rmď5ONG> ;'> u>n M=Ĝ1Ʋ^]rb5qR*Tں/) - >4[i+HB!ݎ"=1f%@HˉNё1.h3#smb*Z.SkLѪokFp*VacIISMjXIfW*FZnY2xF3s{t;&6VO_a{/V^ECvhV-jsC?. M/훞mvvڭkZ_U8 @{+tm`Kx;*[ͣ; _y< {shw]x׽Z^|sbnƝl6ť3gr5pd6m_ûn{V|孷],iJN(%ʋP&s#n00)/n|7ghj^ޝ*u6VYi+iC*:mUKX":1ޞ6R}M<~y%16 %]Ll۲IŪ _v"*֍t8'd U)S]'Sq"` }0LI_J.McH{Sq)>@K( {'ky ncK>kT$kU6 ZV\mI\O#wOZ']< OĔVJ'zF57;,:Ԍ \1FǴf>x3Oo gG~r,Rhe$H灠pʈ\JgrJJJ)z}>eԢh6,M&#d&!Pp+Vg ![^2S 2fQY U7*!hܲ{Jx;Υ+ 7])Ak8 rz/5k+@ԾQeOS7fVFfMӿɗC+KD\3ZK逽gOc t6m_q3]L恺,N\ݾ`{7Xfanxf VmmGUbrN8rzdpEp2PSvJ*|p?89Ǯ]\[)ј1-%F9+&t0,M 3jBddRd! e"uHDŽKF${=$l OoWШ 'Yz?$AR3YM`XwOi\&L_%7}&ly8eXK:U^$Hbcgu'ɔ`kSO[L/B`PQZ@I;7q`G՛(1ߤ~DxdVϻyvɜj?nmԝ=8Z#0b3z~g{ h\VgZ%)lL{썓ɪ֞OqċɴAi@-hQFDd}Օs`,njSOөlNZ#=zAI^"mU}ӄ )L7hP)uJy&*yk+T2>Nc3:Gct B0[rF9 x ^gMQNzbxO,y'W<;Ϡ\-k%#s0mngyEe~"zco=udshyQu%Zz.qƐ`+&c(aE Vё7~ЊzY/g}?v<@A"딅R35&]9d&"XR%A xƷ?stJ:6^hZ-4v>} Z,;~/3Pݖ1~L> %,kHC;"2Q::TeAֹ\^NgQxHB- 1# %N+01)$ȝ7`RfQC wXBz)ZyAz \3e !>vMޙ8B N  !}e{2gdKo[pvJYs?Wc:' #6XBHCP,e_8`5Đ ,QdV$]!GO73?`=#sC2Id ǐ摕R5YxG.5 @T< O1}y?#w4Wy f&K0TS0FXg02IFjwAPMwtӣ8%u۴h]:$ԝ=a8, 5,!;F7oA$~L6]NzNPC`c̙`2aMp55{a0~L TA{2]2‹";/ s[*4 oϘj?&{ **HNV\mgS ;:D/c2i})?,!^$Ӿb(%Q a9Gχiу'\𑖫&ooK2K?3I`4!uRIgp>x+#z5!9ӻ|ZI_ ˏ&o9Ma6_y뛓/uGz[~0 VB;4;5 AZ7՝"ejJ.^-_~9k_ڃ+}>tR|ɢ|}U'-l zv[zuiKgB6IZ]u/JBsk}):;u>W"H]>AP}D ih/2wcJKj:z:&GAk6͈:g#г:Oh;QWz9-.HCEZkt~&h>VNƳnpo U#/>HdUa8^&-tТM Ko76x;]ܱ&o{toqȆs^b7m9ves˥zܿ4#û&)RyA)E@`ϲ 9Ð&l~N,>(:8YTx2Nh%&3IzR$.!):`B(Fn@՜/bSlnær-Al*s.`NlN_ ߦNgm蛱y;8|X`E?ypnθۛ mzM-kOk_!دczeGErںލEWSzBlEmbM2bsS=5ߵѧɸ8x*yjhM)>y~Tj89ju*U{Yo ѯ9yXG<.3qmƞA b]ѰX\)mˆu܂6o;~gxgFG ֙ t r)@Dt Mf̲}n#!vq 8yoSh!(lN#ZzRX;ﴳO'ynA~G->Eh+ziNML[:{!dlӕbuq|o Vj!+ ɂ+qX-KjP NI$Ȋa" ! 1Ʈ_t%aEMӺiZ2l\|v*n͓cCԉou}ϖʔMNOӢ99apz˃Y:0rjU߼pvʇwԼ<|.U=K"9iD3lldf 23KLHo;EFmM&+)Vb%=j*CS} MB͵+@kdL؝v63B1 <wH9*qKtsN;\:ym&;iPhe4͟9b'(qk}RTj:{ZC*Z6̕J! re2!EMM4M&ߎYMt)hlLjݙ8;LƸb jw6;Nuڽ7_7$5& #DBor\ F@^m"kDJ GRtYic YHF uYԉV"<ۉ.gx%a_`p~@dˊݤe7dSqܿYe~*|Z?{WȍJa@'Ȗd6@&1$ִ,y$cI%ɶ$.ݖU,|׏|G7`)NRYt{g3h29d$~~{"ÒOAE%]1x{ghI^4CHLqZoNRv OX\q)=AD"*@ JY+.b0 ɰa`WUgVzzu'>˩)yI|;nkؕ]EbfirBM"%Y*sW۬} dդ9Y̸[t'L0{hD{X6o.tfq4a/Q3ŤJQsLBpIG/yD&`{p#KAn,vtRY2 )q!J0%TK6H0 K(gHz8Wt=»w]J m:oRsP_bmNh/b5n޿gwO9vw lͦCjYS6n;=/ywyvz(6^~|]h~7s= _5r ۦm_6Ou&dj[͟ ; ۵i_:Ꙭw9W%B:Bb1{;|zG$7Yߣ}m||GH{ 4 w˱ ;cbiiqc;l?ud#i A{nr @'BE.V:a } Ȩ1-vA%'NNAJBmmap.~",m`;$ōa$Y{|S09CiY~#-EZ>ryRܙnU;,dy08.hKvL*DQd =2pНFz଄tt';޸GzxƔeG,h%*P ;Ύh[ ]dQ;,Ur4FI@ǃ6'.3%,$T3 { vyb3N+J^T8NI .SQ U^v^Vcj:e_UpPVIeu{[u_0ˊϣ\}M㋯^vՋ/H^^ߥ)Vm>pWCƢ.?dz>nM2u_Є5jO/?S;1kTSi`J9Y~x5, "?kgZuNjRJGX5Kg։djiEM5}5y175՘FG1QxB& cV W>z0ԥh鿅y~,8qw1}Pj b m=Qfc qf g϶P*7-#nI񒠛Kvt2 *t.Qz~-c.&|PD1+w6v"A|ՠ2Rl!~yTߤdK6G3/z5ّ89ىKZ8zk:v? l;;"Z% l&@Jb"V_{~sUYM___5TSȽZ(B8 Q_NJhnG[`'o"ndh˅"u,p} /$~Ѱ6QYDA;gtezo:*jAWՉ,Ȧm׿x5"fA~G|R|B*TB$cQ(sȄ<*u5ʗc !$MdSc\ 9@ɒWL89pL}}fkR:m(sLP )`cD 2A*' *g?-mId`A4Xuē'FHZ=ʏI:O2x\ϬLL=L3)M1 H)A&Htȇ'Ҡ-B,>, !22x42ĞeǸ%#Q };@2"" 1"h̽4ٙtvfw0IWQO.f˳W=iEB̾҈ğ6 NN+Z>㯯^ ՓŽuqyь.th=^E_nYGL-Q:7tG$52wI/E8bR0|͈A,R-Y3I쯇WU+gsgM{6q0â]5!yvD`0\!A#SA%C\..o3ռ BxGj^1xlN?uE㧧|%շ[K8,/T/O9m2⦳j0^g<ӺscF77RXΪoTmYX,JFwwMbE?-ZUWw~llJǗ/]vY`k˰me=7q}7ARtUwqfsmUsU6|Y%-R뢃40JGO=y~Ӆ !`-~FQpA/ q k S̲) UI{=R+>;dm鶩o8Tgd:O]βL[DfAvI6*:z&H ilh%1PcFͨcH{R9ڹJX~:Q2:23ZVF}Vj[Y~yYizPFGd:eaB >!WdQ$Hܭ>p\QJd- aCy)UnAAZ8^лz_S;5ym.߭DŽO(`Y;F;D2cPcXzPhv^0,JP g+CBf Ǡ@J(L@#IbN7`RfȱA+wX\z1ZyA| \sEǂ찅0$uʒ%_FݭﭑoGTjkFo7m]aL eV +XV)uKtٗbHI9d#h6Ilȑ3\^l=$v4Ȫ1dyd%%Klȥf!( %O'> ݟ:g8~fs#U2IƘ*h>cu6 #dIz,>ڔ{^CbYoxvW áC AFQn`[zAvd/u! .̋%vs*1z 6Ɯ X0!LɢCNXSY^?\?][D Ur ie) C A{m*ŅRJ|~` ƨd!x|6<יGNkd T:MwWUXV6Ɇ韪AX{D-JDBXÓI Ov|QG6\~å~ồvzwmˍWzuLqI:7bÞw_Z#6%|_D"J$Qj/N<"HqhxòA$3d~Q$zv#ry۫1z#{:uvY^TMLj/3x@φ˗iqKV3f K\Y9F8fNrm-jW&׼z=zN?TfyuD u=2ǸbtMyWW,z 켿Dq-7Z߫S74!jD#ZehjKiix³l* ͦr:犸mաrlrkR)vڵU7?Ruת Uj漴,?_-MՑ'uh͵>::|9^F~uX6vXqssmGFyz Ug4Qf}{{0ztq8aQΪ#(ם>P5r<Ad *. %BLj>:{TedxpbS9\>UU顋k t4)u@-(;dpHs&0Lтc\+52 Bڴ*7U_ޥf跋1+LXnǕ=d HVǣo r:Rzwdq2` }9n3MkXq,oEQ{:>s/5r?{}Z\Pxwx@ܛ)Sejض7D<@ա{ߑ.}>NcQ1Ksi5aw4o+gD/ ;?qI7FyNwF< ecMg<yl}h:2,2P EqCPÛl0hxtIJSpR2(Q:eB!vΐ58"Cц\;8u+tHX쇵_e|Y_sx"}FD_r㾂c2N>cÕte eCgUi\P(o[׌V,;@,}OY<ญ<;”ՠhfL*ˤ3fǙVFF-omR)i- ܥ } !!:=w*h sX&`Fʩl~1;6 #u>JyxeJŗ6e.ERFf%z)6 u5i@v (7L^D8!!0l685zOs8Yٳ9|(EɾE@$&3@Oe59sd;+(aïiwo&z'xt\I{Dr҂gL1+-d Zg<VښL()Vȵ%dzHբUSSM$\KϹHնmajʶP5[D. EAUv˃NWeq0r/Ox<8>/^NPR(4 a<Xm~kMehElzk'!Fܐf= 1 E)" lR\є+I2I.k%ؙ!ZH$`a7/R<TWm{l.6K].6xvݑvGTf@r{P 'kT1PLov!jP> ɻȝg +@4e$DُܱdR^iaC$ȔiIWX,]Vx2,R`7\*|I\^ b> Ԓ^eJ՞ @4tpT woHaJ5xN3_ |:^S95Ą]]Ddt42^O t<,!j\a5pyؓD+i,,%qw??zwÛ~!p+9(Ic_'f҇_x?pn_4i*@n}|>VsHCQFsCCH}mxk0KA";/ MhFmAXn3 |?=w/&f۲e? fm1Qtw9.jM"ks?r~Mg?9Ӻ3fzM^_͗8PoG%}w6eO=j9[-rZVj9[-gl}k9[m#mrZVj{i-g@-gAUTm9[-gl f5iLWWhj@GbV"l.RmcJT3"e`LL0(We$beC~sj! 'FT)jbIZ.iB4|>:Z*w]JȌKST#,i E%*&*QU`-xo:#MhNZȡLEHCTk]IylrH¤kl0V)#Ck Rp"@+&p !(B޳,,.ɽMiIi>p:9]|'3^j1a9uNd?^2Rl:z֋J*_G.xWfMk|Oظ٧qibہ. 5'%<sO[4NҢNԈɠ:)e;9+d>cm '[ϷsV"i DƿdzgV 6DH $7:ZX?_hm*R@.v;H9vKhUZ]˰; ;R.6{xV!M%q,HfAjy4R*]ЙIɧUR?{OƱ P}\96:Чgj!%;A>k`l )CRPbl ITtWUWuׁu2LwZL"^ˈiP󄴤q͟DC7ܱ\r juauL'e;k%n 6L,\ ٭{̤?6*;լ 9ͽ܈Zp$Eh3*s1x~9Ρuu^ (]9ͣY9jyx;&WZnd^{wwIw;ZP| `o夸[}?>si6x$SM~MM фy&Fh~",yzOp{1Fsɱ`ei[~#-EZcBα*vhIAȘHʔSRu I#oKj~lO%W%iqtקR8וC9 ˇd..gc}+ke)˦IoxZ;&fdep3yR5p!)\RTՋ)c39Ղ:5&.on;àasVY8E⇶Z z\@ךl2XpKj'„" .xAEer uk? H *a REN"dd6rU\'LAc˝8nOޅKVG5'z=m5p'.iGO.ȶJ "l.}]3f~|ןῒ__SV.zo糊jg˹I{]i@}=D%"5``oZX?I hF:O [A5>YHdQ&QDA3tixtp!0SMH׶kUQakEFA~Gx|8 ='$H"uT;Ϭ\aaY4K`J0Sc@kL C ;9cAG14F|tprCOv\''HQ[8-|BBފЮ-h=5ɴXőp.uXE0`cYE."JFc&=j GաCy"HS/!!,u4p Ep ^ s< 9Q$}wFJ'tx??{:6$y"5 4g@uOc=.mK{<6$hJPbql)I%3ÓTUG6tvSgv⸡߆B=޷9krq1- 9,:2 PZ#*Y*#$)+-6qddbAEVm-vTxGDJNc, d 3âҌ;d2!H&H P~?" QvuyYw!a>x-SfhEQ*u$ RT`"2HYжhHgNL4q,Hf .z! #Pp4'S뎌/#= ,^P"CaQt8$19a`ccB,*dzYoPLϲ_jj0Gd' Q_BSb:Qha8 g0]i\Nަ WٶӊE@.th=ΝIė( O[kGA91r}wKxgmWJ(AkSrп `NCvv|f>e>MK6nTξpX Erq1^7Cw 8A;yPPשERӡm:D;zD]ESItV=JƷndo^M_NoeݧL~U2O`aXJJծ^.u;(_O_%ίĬ4xYLnNèM`E?nd ϟliI׷'|Tqgʚk.lӍ 4^v>RtY&e`, ̲˕LoR_5C6)7YQ)$k,n dӎW,@~:R,YYd3YWU;صO5rZ07oՒ'o;jJ(:j*X}@/?;Y s^J;zԃ=vE? @Rx拓7F{0:rƌ7)- Pq &OJ9{Oq2*j!W!nxKr!""^P̽ScG90.1T3!ֳE|2WR4781HSO)i _jy\(Mܽseu]Y04<0smC `x:"$r/6 GEو[a9fh cq(Hث`E*L<:SDcKDpC,s0M.8ҳ` Ь$Ix椻T4yv6Z%;oZM^x:O0# eV!$# K! k9I9`s6H3 >QBѡGL(z:EQ|jTz0vV46 J,YPL310G#ARďɓuƉ/Vp:q̒O8JZ"x惕("4PlіPh hFb۪E8$ehRRFv;m=ݑI_C'n7"韺٨Nhj0jVos1"eTL2]g"JrlQ?\?](%2T0"#M%M^L$Q"F Vc9x|vz. ?3n]n$'9Iu=%Da ]\Fa|&}w~v:N`e*؊d "$u>bgסrP:&CY\7lI~L'Fuh_;ua-lڟuONneGzk184°<_)TCۆ$uXh8;3)El{4i?Ȁ0_|YvOV8>$]!fQh`ke]1q#Qݴc0ՀΑǑud~;"ihM?Sѡ?O3p30k$Mm/BR&1LQ=?rS>z-~5 7I: %Zdye6di cnjdzC,\^^_t!R/6"@m'Q>]e:Hä_mNFDB5=ȳnڻz]x%Ϊ?xݦ]ֵW z*O !32QTӴ.ī)= ƍ7H~%jSe&-o*{ۀ9PVnM?S4/]e Q ːf{9\]%Wbս᪸AaRzavm!9N6򠷅]>? Q;w[9(mq|M>6)Sk`J0f;rն؀Q=ߎ M[Mnlp(b[tjZm'ohF{6~7m &؜@ֱv{`w}e7!nq7ȵZ*$y]*[Js(7[:R9vA"e i̠)jI-ND|/yOϋ'W>sYp;:RZjkm#Gv W7-p[,v~9 `{HIN&baђ,Qt$bUůXůH\tl^$%ADKNXk}FIW$0YuU}غ=@mչrOa=Gi7?F/Vg7C{Z.H׶:E~oJck w:{1mfܹ^Q0J\T-^:?+| 7~%O>\.o=}}45c^tZՐV# CV<+uu:{>Yo 939X";y\xg4*Fn56TbG!z% gc^Ĝ%i+풭7t76{AK|FY.6 `t`A!Kk ElAMp4`U5882ZN'ǹvs$x*EA]Das!TғR|e 6(zz<gڋկ _pȠçh ~EY>g c'(Pry=+*4mK40 d4Cp5J66e W!ޛ<$b`NUĜ*9YUxHA,0P*z7 A ouf%## $-K3R*)<./K0h00kTvu 2aG4?pB$[g*D"Cg>F/NTs y\e|}:W yᭀƒdPeƃG?^:ye)3Ks,aF:nώEaJIkjPf\&e1gZ̲n°IɧUpRDcHH\d3;BmPpvKh*{ٲf#Qc97Q& YƲ`YQDMR׭+ к4rM] @YSL^8IIJZK0jו5튚>NZO3^lK597_Ih6z=2%ӷ%'rmO R;Γ=bʅԴۃ :Ix$L`<R LҬA&:s\ȄedBIkkc\!U VO(Ox#xZzLFe6햱Vf IƾPWGg"YqHn 28&6FNN"a?opex9r Pu |g֐tsI5*0T˺+3$MMЀ.L. Mt)`lhDl85\v58{#؍T%$1YJ;"2oFjcdx.=J!2@^tɐ-Ap$U*a*@Nu2v:\lSd(\,b5"$>:Fu> It.2B9 (E``A 溕\I7N $wY.B:Bb ^!ʩp[ĶG%9uV}lmo{1\hEw);9+&\.uBMQFAN:3c$1jq.g`¶Ήr5rn m$>xQw&ݣ6'~ I)r86i@A)|9 }!G@ԩSQ';BΏ"U+EusȄd\ּ&4!dQxau|Ԡd'e5u\s\˷?ro9./!7D  @wӱJxvߤ9_7g`4ǡV75Y#Ӎ1x3!+sqw"EۣA&ճ[) IHRhUlr=0WCTq[?AYt X VH`"ŴX@}tJOg˺O,RvY6㮟]v5Ʉ4I4 YlQkW jGV&y8X&W1Eړ3GKB.~-LCr\><,=A{Qx\ @ "&гț~f&[!KvenyIsZm!m)FiZI%5ZdDse,̉[ltCa{ߘ{9H1 k{s$ iB-`pAZ}k0r=raz㉣'r`ĸ'/-S@f6 Bnxbи4NEZK alm.2SY;|KNj{|h+\>zYh9)`bW;OQ2v*Io4$!2 +42c sE8E(|uozp:v: T Vbȑaf Qq `*&=Hޮ½=ˆdYλ2;㏥{kmؾ=]Kc?ZG`(RjDE%죧&Ech6Is`cxƓEMO-k<Id TRK/Yx.5@d<ʍ ? ǻg/^iALe0T#Y$#O;;:N!ҷzZtp(`h۠j{v}Y /0EvAygm<7O76Owvld]:Ta޿ ~8(}ؽUBx4۩gCcet'5rKZB YFӛI޷h~Dn-BI4N{oȘLk_,S ivs~oۻw^dgɻ4" Rgt `77wp4y[fa,}e׮Z $쩯=edƴ# III/ ˇMɢG7X8+vlɒe5dbW=UF%6|lߖ@ŹPu;B,ӣGmiiċqhmVD YQӪ ^/m:ZH脔4=7R ч3;{Å4Φ)"mܿ'c4ȪOdvtJD`z&X .Lac]6ݫ_8/gZJC6O9y@xcmbf7#4%2f?zO[ۛXtlů-}W L=5?zy;:fBxoO7剟t:/:kڱ>H]47dMmErb&m-,vt)LŇ ?K{!oWUޞX=?l?9Pew %LI)eI%$ΙWo I{^ʎ1zvݶD z/L ij=}uPQLua絒ƣXryiYx{=s2H́a93%X3䘼r{y\qw謋{~'ax 쵄 YÝq-5ێqU-XT廎Sco;~:e|f& BJ*aB`P!(ѳd3c-{= -J=9ޠxP*(%#89 YZ!t4Hrdг&f ^^ڼzg_{ԻxSW-zbYgi^"\eQI CpM&AEojZd2QLUA77s rGyZ~'pGg)NŸ%6;\V)@Ddٵ{\RkNOC0u(iR>+nʆdEJ?ae2қ8M+KX> k}; E=ސ}%zC> nz>7k/{A!aJJښ^+d ;HH50ɶKj[2ٻLi#[e=b}1[x[ߺ\IlJ:Du"+ȉtUe@hi-~dzi=9@e3QWMaQ!"Α!^Ӱ4,#L7e*)pXQO ult;fƍ ' A40g@׉ɽ퐎]Oz _}1T/סH/"1}?{zm( &`՞M[j,mؿ6M'T.d

qF5òtEE^n a{&Ҟ<L)x.S G)JtFm=m/7 IQXZ-sT%a qH*҃* ALI P%QI|YDQ,! 0$\FA#;|}8X[)i '/-NWܑISÃ,Dp#8:jIp#tIU\uhvn]jn)XgWpiԯ)(y<,Rgi^/+2Ш'=z$ mFb} y^`uZv`ƿ6u/ew vTIǹc[)rO ѴR?ۀ{I۴ƍT!3FRE|Fŧ[XmtzI'`2‹g7AZ5x@ .`x_Z)5ͮiF?] BFǷюx5m v|ׅIu}{~:f~.Ip<>/3hs6=awuzkwdEeskQ5F*{uRpj.kPKu̬vuƍi3reoF=>.,z%pl9 ۚh-ZyۅHe ;0wO|~w(Y@iՇdvNf}|Iy5J=A[zV^Y|$ r~g 'rB>k V`:- {Ͻq2Fn]< A2!hc2HStyqT11e+\*HsBa~S;U9X~*W<"1o^GPgH˰j2!H0icbJڃsRL(Y{~G1:bt褟)B#ړ_DT)*S̲pErQi{ItTH3_"y EOgٙuP zI'y,i0RY3s0-n-q dz-܏iGG/TGOp =#ZIcP΄ N{pG/ ""g{J3ˍ%l LH*&`R KwA%x1XxZrfJ={MRw|䜒QyG8mԌTk yJW&AZLF <앷+qhi~VO&W6%{vLSظ|#"gR{߃C3-}HbZ`ɼlqefr9d$}Ui#Gf(OAZMNCURL}}H)7d^ ]֯j9K;TsKA c7AD4ZU}HJkd9 ) Rw{qg 1T[:n廽mJ34v~w}!'O:W u`XѠn?O S+\O(69^+vd1tf< %:]z D=H ๋<{- 3J e|. rʞ3Ff{ź1)(. [C !B.u&o39q AnoliwAA=8xGe8i r\{yp"It\i.WXDzWcHferxQYZ ܐLY)!#@Jb9ENZbQ}YQ #XJ.i=ޙC2VnGg<*lI,]I]!UJUdwlÕfx2k\?^6QG<䒎 zP0fU.2AKQ1%&$^-q6J 9 %dzHբӖS 0XL B$bߖ7q6[JoP[hzzo $*vyU`19gr=]FoM`r-vFP( $ 5`!s,P{ kGc/AjYQ7g|_&CR)Fq$2З`3r0>"kLjI/YgH9ٔIKpU6܎x*IfCxWRcAXP[P`=5*&#) R%3"qZ)6͐@PD#s,vC 9q"@bG]'n=wMjytݚ?NRS*6SUBA騚cjxyU4{[ۼߏ:\}Soi' +'; o,<,2,mP+5Ԗ"+aN {CR}Je rK6WZR^NgOWOWSD=(W /R ˬXө< $ 4Y(^Tfrhee3ʍg2a.mc6RSN7:)x_2@Mt2 n'dpY_J[\ x֚D}8oJ\(q>݇[}fgXK`ZJbN| UJ8$4պ*9L&@U}i±H/kN6*˂'z'zTsYȲ&&hܕBfO)]p♟g­o]qCRLѓb帰b.帚QKk7*!MWejиFQz_Gs6OK#- -Q60ZoQC0v<(L}7-Ѳp`$%lCqd\|2<i7 (5V7mtͥ\ C/V"hX\r>mvvGbΏZlJW-ŻU;d4m~M]I*ѹµ=gW}>I59$<[3(Kod9ȼN ++aOV>}ri"Q06ø}JH sj.Nb KrPLjι.溔Iu9&>Ȳ&[}YjfHfc*M~ m=$pը)I[:hhGɖ]KN+>\]wKS $CvI^.q^"6 *.'% 9/5rٺ1R7tdwVay-k+,-Is/:y956cl>EPSqĺ <ҳxSEF_eOlmoi(M5Kk5,Kٚ1~S>\ut$4ˑYW+\+R7ZZPUϪOH>/TsS2_qgoΫ-h󟓛<-Vܼ7C{]MuQ}"R[iVZ@%g?!^]좊wbee6&>7d g!Ssy.SntzRJ>\]MCaUw540rh9򃻮x<;ć/'g<'sp@ʊg-Mj(} S/@Ǡ]N0@oߝoe<; 9po- co[=uk7)JTtu6&}oC͍ow_24MP װ|o- G"}5кo?6 du^~S޴$ܷ2 lIBzH"oH`9pTKxi](Fgr[{k`{P>laė7~wBkm_rԽ%SFRsd%s$yy9Ҍ_SP#*-dlnR4Km;YzQnɜl%>݁LhAC4ZxoS1>>]xXѦYrΡ#6sֽ ^iiIOlYMiHU}>zi4@@$4heX8}#x:-5! }$)nnZ8_+XkN12ݗui\VvFXp'Կ_"š )'eJJ%ȹ.8k >k/y W &P"4x>//o~7tϨ{~+j"b&%+ۙ 0 ^ʭ8"Oٔ>hf};lxxBs¤5e'O9wiw_ۢUR-.:Ὧw 2BpPi25k_^]5{ @{ @*ʿ :{Zi-X@t-L !oo{G_&Ɠ2Zqw_eR/o(wcY',7|`$_{pW;;S^t3YEW,;@ν)<]wՎr Kmj8 QunnyD:E̋ᥔ+nC-86صFª:#cw=B4'V1T-? ocdL2<:I#Ty+'>Q JfeU:.1Mkyiӟ>"1'8CνoMEI])iWOQL Xv4biVъdא[h⭓Em^yI\Pi(sHV̵M2"dl9tԪv[<™lw3R2ͭg6cW 5e=p8ExaQ3L!}E!ИU9ج cV2rqc*x!X Ċ=AI+! 1k(ix/lPomD KJ)-*PMBKYKIs$!G#M2drRhy۴^:2j͐x,uV%K)|Qp&Yq AA׌as@\d!%ʇ@HB_ x ,\&VS| ђ(:|Ab!p#H F2Qhg7f6Pbzb TB#DgK4#n[ L/=Ԫ IX#ɐ [ΑEFG@kE0O 8W/ep;S70 K p§(Ry*0X8T9 Y8Fʄ[| e jfC*^6+:jr>}p' 0"NaJB41{OPfJg3wt)A SĨeB(Z6CO-jd2rܕSb$ "@2aawPiBBh!XȠ>hL21GPlc"gbSX 2zl\Ri͠j`oȻL >"~H0N`HSx+ !R!:ns8`P16\:A !I1Τ0"*% na%n `L pP%YI !PQ".0:EVWbP ^ 4?#|!姘PSᑐjV~!RXkb:KrƸe] %XBT#cd%婚% /*'tRD A ِ:DYΔxD#Jޮ 0$U@֊lRP]E7J!X fJⳊX,Y;AX iXt,+V?%uddIѵXf"\WXkp`c#.yջ߁b<IrC* B.Zo2@9:#D(/!4Jhpl2ZepyP9@G@!th%*X=(f@c|NJF΃pʼnI8+s-, $fSV&J}d@BPp;қjR|<yf!}P8+B>6J4Z9>hay!gѝ`ZdJ ׀,yHmJU*s]"K{"he),U 0?'D|#&%JXY1E]B֮m1in=[N tFu0?{ئդBL.Qƒ)8鬅t 7ɐl ѱ`0 ؾOm;FŨM$QZ\;] ގrtݣ-7f섍ܔ9n0J$6C謏ȇpCaa)6\#nF(\0wzB =RT"X( 7B)ЃV#6dXo >W'3dE\" W1qgm\!ON.5;96!I^hyİ2aF9_V((iQ@a!T>)<$#660 ; |[sԏA!c6̋ѲdG)[l),3HkШT`g1{ `He$)0iCB"Pdc| ޵|w6#;erXQ!DueMG T)GD>(F -,h#kA: dDj]/C_DYbF8|g,c@ؔC&ᣐX28b$ .2 fw jct3κ|U>~"" lph~nBFCbɡ:.kHҲFt^;nꊄ S4-~W+@kGիR IgS`TN&+,4zx6]R_zfmUMZz;vG\ h>,mG)+^r֕ffst/EmD:>hf6?K7ӇQ9`B]YgU_ø|ePkfj~`U R>-&aqd@VGR%b92KL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&e5@/L=&0 .1# 1b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1/@SbAF?&PY~:ge@rX(b=#&b"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b=[&Ek)1y2L"z2gig)CLgSKL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&a]/7͍݊W?Kiy뾼>m/ _{b7l<*.% :`G)@ Y֝NW/ZRȽC%F;\~ަ&gӐj]†g|+I%Fqe76HnHvbm[`~h]>'I@P N]vxhz{7S\\?amtv<;^>oߦj])|Sߦq5}p9U,a Ńmtrc~>6?ߍ'/{֊xY65WY7TD|Е}~FŻ?2΅'g'屛wR>ޮ9|驪ӳaa(YF!op?o5(םǗzMϖtэާ7//.%/¦PG@-\Q_Q5^6n"+DODR^&PF^;jFDWZyjP+νM2zbPvWA" (,3ƌMv$$w‰L%9$i")&)cJo9W줎^3X!'&87}, qW76+BÅ>./q4ys#͗K_/g =1̉ $˖eA! WgPѡlzP,7MA@S ẻ2d,L B0ѺEA"} W^6Nkz(uSwiSE/\cRUʡ6\ƺm3)fŽ*i ֠SQI<Ԕ1/] xd#U< y0=^6 V4 /tveJ\5T24LB{'Ώ\?U6eh)yF*m0,ta1*0^+dXr5ܸ|"`eZGĴĴܾ^5xa[OgږmSː"4"G΢NzA 8jĹe!q60bccϡ1BJ$:NiirNF|bL*y]h}w)vA'UY9{[4Pաzqѡկ?_]W5s_B8^VIݮ>ת^ItZ8(WS⋯_\i7(xQx,}HZx:9MR#^K p~;ŏ^40|']?).7~R=!}t<Lk+^b*F,8 /x_ת7@]c Fɍc79ɺj1"ޥ󃾕ˇb=M TWk&S@1)1WW.l4wfqK4ǃ`իn|a ;BU .[Ҭnڢ,'Zg8Gvš{4*u [ pTBK'⠛!99HǶ^\[+ʱ9!5y-%o&LvVPQϺo~+1=›+w4K򋾾\h8:,ͯ"/8]JvS8`LI;@b$O!bë^ ~Ug^kW뱞_=Jn(Q-o"ڎ{`Gߐ,fadxq?(G zPdK`+V_*uÕo1z*+X ]+ Z>s? ]R;öO;Joנu4,E! m]VEmxUR%(cD2L&"uhRID9k؅Q`,)- >e',Y7Xʣ9 |=.)lWu?nc"(kMcMNwx#}(6ga%cTDŽeבIGdc|Aۨ!koUV,#B-$r,$/:вp!:YmYBt>2ʤxi l Mp(X6RIZVNj֤V*} >1l! & -K6441܄ʛy_uMK-3OJ4EklU9捵^[͔j~(mX對dr5@:/:1J&c tbe$t<CcBS<0lgIa:ķY/޷7?[V#ghM'0&q''Ph#prV\y4X̫''5/:Dϔz\6ᣬ:(1] 5Ԩ\ C=g  az-1m^Ty)BrUՃS{3/y\nlX&U2nHO]3%d2ȭT{:iZ1Ŏw jL\ 1qǯ[5ȿOx<}t[~zӮ_iyկ/&,^]v,R97Z5oN'_ߞ j uWׯѶƛ8|'o]7v|r2O|2᧩}%mai 0l ўlo ޺յFl[Y;uߠGf\9{K|VzraV5ɍݚܚ0M+ړuލB}S}ܚ2qLfn2B ~ _ok|P-lO48\+|΋Y,;8y_o]F;{h >!bRta8 DlPmE' d@=dւd݌/Ǹxrd<'J:/JEũCAHri%OH9 )Z7OٟZXV4QCw)ik.ej = Iq"'Wf^tJNΘZ"k^ $hN٨v4[ѧ]eOV^.G}HﯥŸ P@Ĉ ON$1E4`a؜ۋ/4ozۈҷ,&[ ')[ďS1Edn[iprG~i]{G$x+7C1{#4PC@^oMUI_YYu;yqnX̖mo8U ~lzRZ+~l?Q~P% K /g {.DҞp'1$.lz?9w&n&c -aE?>?P9;a|՟{$>^n ^o<[B^ PV y'{ZDNIJΠ?;DBHR3#ĩ`Vx!r_=  *2.IFŨ4C'蕗Tc`d2q(0P-W!F!Hc.#/ 6fFəT{*1 <3 003+>cBuDy9\+7%zvفx`vްE4Yw50<uVYLT!{\aJiG #8WK`J:@KGqF 8B ܐ]z v$ЫFd}\ 0%iy Ov]~_㟛#7?-\ !JYM953$9pt@rTi7k10fop3Ϩ;9E#o`_FJxI/t5{hA٠,wZ !%bFh hph4ZR'F;P.- #; I=' e9BMni%Fz!R֧:}S6boܴj ^gUSUn8ۦ0B+Cy֔ AGogˋE6|U\/WeoC4IԶeBz|nr_o8U$T|_1F,/%e;nMV+xq!C\ivZZucYhuU!S۰v%um3}!u-*'կmqxמ+0tEGR7~ c\ ˀ^S;G`So`s% 6-/=Vv_ r1v<ݷ JHGk""[{md>51W>;1[s`:$_|}r>Cp1(5x.U RȣRFtJ͡c%TIn+fK(>KoymW#Խm͠[4#{pvk&͕^0e|̃A<,ְJ_uunzC"Clbu-,ۛB3-μg ^ D1$/AuhEGpT&o ,ZGp X f85c2(pP@\\̜#F?|JCU֟} - 4wcdD%ퟜZnYh2jD9_\jo~v&cK_?R~4~4@d( DH!dt&ojt5NDtB\ ]RRt :$IsM$B#zj3 15ϴ҄Ҽbw1Z0_kv>xG;_"q'y"=j/_ʶt}ݴIAhUoi,P-#=XQwe"40F{\K'Z:˻Zw wyKMɒt.nj e#3I{A&*DI̓≗% kdb&k,P <)PHV'SͷYLv s˵}4z`ĩ(̖}Z>pO.EMtZRoX`)KLPg6$Ya!)WFǙek q޷S# ـ$i)c#Nqh%7)/ךPZ)sT8]4^Ͽ 3YF֥\``:"$HpjVW*Q(70LQdTF''FSЃC-},::i*(%X̜XW)zơX cH6;烗]~^0roOt:8,WAd>yw\:.:G0F#% %4h' a/32lr^:alGM &aU HЦnD)s#Vy,]w jGvTNU#)C'D\H(ij^'O%ާg7)0ò`|V܌\R_/Yqpy zf(QBjh-4z[T:e|QVWsS%+4WjIdTT~&l$72h;x\oM"f 5Zd5C7ߝG46),P`S, gţȩ r>WC6ŹGe m'Tnig RGSJmΌ6> D_@L-eWsy $g\Kr1{#S!ғ)|<9&-9{I^,?nf/DHE$ < /2ަbJW*6TP^{fl+aG|CTc{RY~/˸Z[c*-=9(h?m9+˿_uJR\ rƂϭe>w/iLDmskD-YOf|ݥM&V.m6/-UH91,Ճq66,Uqe#"sփӃtA6>\۬ۈO%c=G{rb)t&$ I"ɋ#НS-Y`HTK(}R~c3{xr 8gn3[ )%X SlBexvCl CUT; gdy1zV36e ތO!™M _GYPOܪb-30|B c]e)E:xՅAd\Ea`ʍ(sԎl8T p8ړl8ێeujhJ$`)]Y3(>0@Y1VGCD͑QlB OLe4U (,ða8z|ש;N=3m/)=ze< 7dr0.\QpFՏHVս u׭jQEA4\T"O^1(4w0 0pw[CNX B K !!Gʍ^ywۨDp6y쓰 J5e`AiҚD?"SXi2$u` ';u wJځtXiߟy{2Lpf{Ok&0&sIwVrVQSohuB:~N+YƫQiejv8-e]g5roE jOkw<2Cz2l <9p6K@5*WB8- \f;wˇpWC_x8]l 풩C ݰȥ{Uhh|=Z8\ֳ7om卭kд>Gc8t\Xrf*=Pdު~rH뙐L'#:ĨYL5T6`-e` 0GM,+&9NxQ >^Iv*u벬 L.9v<2){޴="s &+フ1M훖z޴ 9ZpbYITBv[Gđ1ƃ4b&:=/t(joF6IpxaKm}J!,E N&x;>/l8-kBNbaƩӈyC=C6 Dc`,YN8oNj?۬61̱5Ň, .C2 YR&wY>~^۫Mh|H/dN)@fMs[vg0 ϣ2]}mW/]4/gospYSr<WT8y}.q慟텹:nJWݘ X oFz#<%!Ҹp O\)A2 ˇ$d.2gecu)ke瀔<|2#_W脢"G}̊ Co~|]afB\RGo @ 1^ R3T:5&Κo/"?/bUrƶ{wBp}<ȒiGP=\zWl BE/; [GGPTvxsTL8 Cך H "uKd6pl6pKkt>sUGc`_'Jc,Y1VVWH\#.`-JUȹ _eJ\Qo?=9_bpn_մٜ+pQtqGIKAA$; N3J' DROO|2lISXOxa}R֩ʤ a=CCWIN*?%AVmYz?lIȪU(eZJt {$v:qIgx,hȒ$RD0|&>vK 21Pǂ")@bhB1b*L"}ʴ+.w|BEт/p(7*txpH :qK"Rk?"JFc&=j(}͇XL+Ц^FCzaAK2A) {8\ "f{D{#4*OE:m5HA$bL99H/ѱ!1QQ9ruiS\'b/m9PSFJ:lghc I68mzPCT$:Fw3mڧ{ܷ \m3!kd\k {۔+=AEVvf#)9̰4N+ܣA$ ҁTI8Ps].6/le H J.AJп,\DF `~l}:}!Q9"3X1j#X腴60i\{O$ԭ m/H9ƐhX`!e$$Mh vNX9C!SGZ>2WM?O\|Ήc+<܎(G{$Uz]rDJw&ݲBǣ"un%($qjCߟp٤(=?4%#fdF5W)d?pG'!;O;L>Ds5f>VMf[6qPL_c 򕋪Ӫ1Sm hd&`݆cNM/.o38 1)_z #KGY/or& [{=. ٻ)g&շƓwβAQ\ݻL~a|>E?Ւl8_Z3 kCbkuغ^Cϕ;@GĊruuT&ـ-k^ *:٠zÛl<ޝ% Yq+h`fRBA`ݼQeϲ'˘*.D2P.ѳU9oj6|!P T`}?;a׻6~&=bI/'*xcW:#gxÙ 򤔁{ꍓRy)=͓= c)X.DD w*pS80&O 匢EȁqkӟOur>U'hI=2擞"݆U} QYQ!ʖv]i04<0smC:6:1#'[C֟^SRFY7Eޕ  ^e7r\֣R  5x+FpKE$A:mާDJr 1]GL눢hk 0aRi]1ax=fԃ2ҹj՗+$ y*PFގNK+jq&n5RRS޸(X ToՏ,e:U]vzhReL2<>W? UEkjwl>a\ 3j*D}4CE^Xv?' ]_+w)HrU:fG'V$JBv<|pX}S:P'ZZfBබWMb.+Hl#kږzXezơqW!T?6=񓝄\ &2;\$B;ܸ 41/v&h^FR?qI?v{&3ELQwvd:swDYo;6st9.g n7GOFCE:tPjƆ_=76n1O,vyslCZ[{]OY%0%Ch[Μ:m47Asau&Lr]R99X 8d) .3,Y1k~~A sکWhTFx#HxJ"8&P;'\&n[?lsPT bc,[QMN肼TƬž-C)GH4D?QH%VXi"hai)<1ԉ <0b8y@]zށxrg[\b(A 9׾PRj"ȝ "tJh0JъK1Lmh ],l$vT}+=b2rWCz!CeO֛֓a#Q^.zVeP$ot9o_Ců5^;i@syj^Mhx~UWadϫׯ?Of75Yա_qS+^X}m 8"m6^L'1+@I(jF k9rWaԐ[{M''NHJEe)^G@U HL9b)% hɔY$"VuLd̹.vCu\ԷfHX]Jl_t1IQqn"$gq9Ƣ 2oHPM wx\y/Pvh(Xj#PI."ht*Ř[T-ٯgdztꫬ}uoxǥǦhv`?Ѥ-xܤEVym/qVx" XA (^%ƹ$y3XpQ_=v~he>zhHO@`uUH5J 8  >A%\\OXύ$!\23 L$!:R:93%ᩴ[M1qm:%Z;/Ó8Br~%'J$8P>RO6Umhg_;n"8=y_q(~!q"ꙵBNk-.˻Z'dH`*ycZ` M`:)GyGB P"^[mR$s jb (w[BD Dy Ԣ[#Tİ8w\K 'CW}KJagp ])(*U:Lb@JdDIL2X-[k yHG%j@2dISy8()'Sh_Z)q+j0kXͤ/7aglE-,AyBS> IiONxtR&,:ƭ.R,9!qzPռ cƜ&AR-lB*ُJ1,,62BY ŀGOJ̸'ʮ=$ vy;;mPxqFbGďg?C\/ZgQr(.¸\pqvݠv{#$D"@4IYcEL c5 qx \<<,6:Naxxۓ3̑{Wmo9qh>; ޏO~l|׷g_ 8AƄMD֒\M(X! C@ģ#PՂD$nr'e#bSKkV" b ArY,YHfUY^c|%/b&?UMG\xLheU+*86/` R"Sb"DetuƢȫR6ΐ4mHvIID&Gʈ h j4cr\ 4HTɔ&& IӆiL|$N9d6?p6'[V~㙁9I䌴T:qFQA@ΤO \3 f`3pźWֱ9WTulm ٮSy0&8vФ|z5BP/fT~Ł%wN,p>P{0b_G]v,Ih\n5/-Y˵d 4qO$с>i@ "NklgP~X*@6nlJJDJ2mAh}~|oMn:/ m b5PDQ 4'O@GIj0$(%"qFTDQ (`he&P3X(@7hr Ey ӂg15xR_0p,...j}6}O.T~%* ~&.*Y5vg.>< $|"ir&ܙy(LyߣB: m4:_-N&E_Ov?W b+^Lg&G)*gj&ؿ~ۃWäd7CԋF2I8JRm. V8FQx/;]Z{/.Ic :`֞Bv=5ڐ]jy;"4#H2](_ Vq RiE@ +1=Ub {'0O4v/uh;۴V( I&jeW+cAme±&6V:D%`C+GRgx!cq]vl t[ bf2bP23 qNc@+u z^E[PB]Q[=ಬR uWh%Wk1u2ڀPi[98f`&bD`0ciƗBC-ά :P%@ ic-. `I. pqT AP{ TtgB@Q"!\`#);9/ͬ=;#v5UuAElDT`3/_HH&9%R,qPBi5 YV+GGg`niXEjE},B4 IBufp;ܠirngU/f"ҊY7ICQŘ4=LXJB /ڄ{aV9 LpuFt^ւ[3[Xw!t$MQ 3 o b9:pĥ8@XDJ7#+J$W{HT*#d`!&SQd֣2:/3(Z RI"i Yk2a(ichx }Y%H Vx$i! ox6cQY,TGW>/A_ż d-Š.kI|ӧU.}t۝qr}[. O&!z*4ʘ} ]{ #=,{ԥt9(P"Q _C݅Z:$ճDTeP PJΚ%l-ѧ@hg]+y r0]BZx _3D[vVhň޲N+}ԏ.( g@ @Ae#kvpqAkOl: EBiv%&HTT@PqkU[E¬2,,Hc;&gو.BPDUbhGu+it֞Ew64IxTF+Ѐʬ$޲Bti[ q/GT~ 42YIX6 ئQ}A_AK0A7lmRj6n;lE^QH"NdviP'7 d!g0Eʟu7/'Z@ &eCE5Ff,(*Fb1;L ;CTXD t أt%i#*tQc@sjoN낙VjtҢƚ5A&EJIGfݼLFALFzfj)څ'k=Dw *U&f֫YW98E6V@P8U8P#,]QZ4j䋞P!V&Ex4CS"p#fD9YJD9k916fTS.*w3.:0z(#\jҪ\ >$#4Pb鹓:k|A9oU KW5ZCPޙGD'eAA(E\^ܼZqNҀڼ 0dP[eo{~}0ˊ4н!'c LW]Bl8 {pH H -Pm'CR>Uc'JmIN'@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L}$2"DAt'D\m kI X@" Q9&bI &bI &bI &bI &bI &bI &bI &bI &bI &bI &bI &tI # OQ>O2~$ԂIOB$@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L}$иPH !,CYڃѓ@H&>E-0 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $'D}z^=vI[M~7ZlGܕ'qGKؠچɃu|^]o|f=Ϯ/@_[YuI6BZڠ|sto)kT'))E"Bweq96Q]z:8E4}lga$/_O7[Tb/yˏc㚈*h<~$NW?+[t{%~-k*b.L eQ}Q6ǗW4lVKwޤg'^(i;8ObQR_ʒUSH[32y-p XW:nf5DgwC޼[?Fg׋\ǹ ^XC{$Q9z8ut);8~KmYVf_{kn P+ؾx4ˆ'\]mOf% ߃NKT_Lڏ߽^,/NN7^·:hd^jˣs|5g󣣛F'qr toOgoGѵ#:‹ٽT-Co|xs_-9޸W'g^ "\_,7@n rM'(-rقOaOn.Oso5hSPK`,=h}R8^8\~9%c5<Ӆ8jAjhQ~0ٴ!X#eݵ`;lX5ZrRWST <\Oey}C!qly+3z1g M@8&ߍ{0.To-k%uX̴rSZ=Rq'漡.]\> J>Z_Kz[6{Y\]Jh썷/kDu](/ o×{ް{|OS|cftTW?ԍ\I螽Fp-5o=y^YbR0[/bX~ jz.7[p:?A"Zeu0&!l-h;xYTeJTh?6Tսбh0# D.mT{\N%%RukF:9wKUDFV6 s3}[g_t&Z\,1?%86;>zcX)lqVH CQ UW7-;ߛ(({}O"+>QP[+L>̧& жgT ɱ9(Bjf0(9m͊NEϪ8fF89wěcqA"q \k/:BgdD'YK32B Uʒ;dd6su#ԨXC|bZ膘j@kQx%`>:շ.2a"N"vc@ONO O> "9?h{~rw} ;L)յwlXeKSi{{d¬s@!W!>td~/>n4I|JK޵6r#X0Kq&C;k #jX_zXvxZn6HV"UȚuBRpޤ%cRQǦ;=E&KM/7) E M~Ϧ+Vh ln{/l?ƛN>V lgTۇ㏻7̪Kfբk]GarJ ϖ]#س%YڥͽIZU{$kwթw ?UՓ;9ڟJuCwvެ*9nݿ]ݻVݕ7yn~F;k^y.BK֏Qq=ou.nVzm9ko!Z(y÷lxCvHkkEj3$:j`;6a:F4W&fٵ:g dZ*s,w_| W_@<ߦuimm[}{ڄU^F[ȇA]R鮀7R& 1&k"QK?+&\Ȍ:vF.x&9R&.@ۛ0dP>(LhZ[FAb0:e y,![6N$)/ L2޲yQn{vӤ ZGyJӓLn1qV{t3@iZ+IGq]7l xׇjw ~rXUn[w{={&$fݟ6 $i>u톤0>FDOgpΨì5lmGiHaO7 ͮC|/6꺶M2K^'Om+z,KTlanR$Ben/aa:&9-rh 9e$G,s kmmΆ%pxS;|Kh>Es}l{LG2"ab/biM[w=21L5VP;~b>6fHE +Ւ k‘1)) .p8=Fx u Pް-% #5 A)ıH1W ]uEΞ;k :: &ƩӈyC=C6 piYds0QJ)RuLΪPU3&,w٫W ,/ן?LM1:+S<Èr9Wcd~Nߋ:oR^U׫r٫7%^e.kCpmƣ-dysOӺm76eO>kj~'Em5˱qN79f84?8! Kz0LEeT{%@2c$y7S }N(*QwY.?jiqBP/*5p!.ʓ b1* 1ΆY%q jE kFw'A105}o e\_d B&$yhmT0m[ FvD ^v(;L1:',õ~<)`H\Λ{_5.}o9Yԝ Vo$!"ɜplVNq6Q ,.p@V?}>hגWL%%hڲ;qm)[JND.]JaC\'_Ѐ?kV.&+fӊkgYW%i_%Z{,I-,\ހ v 5 DΙn$34%O`ֽs SmcT"KU&UВ ][%9Rοud9@z^B(p {NHEv4Z% Y² K`J0S>Aw1% %cAG 14FrtprCON NgRK Vn[jW?[U RYU4c 6UDj r`)d4&jciw qQ`: V_UmeT1H j]9 HQ0{0hd@1#ؑ'"}Co h+!F7i#Q)a6 ?;6$y"5 4gй),(T~3H 鱜Ty -_ƶ't: Oڤ 2D$e:WͫgggPyQQ{"U 7ph=:YG!/Q:+sGǐ? \ƬvļQwQbF^)\;E20IoZ_}̻l㠜~[0/yl"mPC3 ?7Y}r^Wr6/yGUjԢMTO45?UYnESzr>HAֆiٛ-"/G7.AYuP*Q'_0#R%Ej e;ʾFoX|` .FMy?R;1^QO?94Xwφؚk>PlӍ;sk}[wEe6(v^rQx1YhT%ijz=YQ.Mn lWY(M묛0 JR8YL1  vmk $TQOUhuRGΘ3A`*NIy.1 'Vy%=q \S"BDD{ǎ:s1a\bHt"z!Ӧߟd Y>xbe}-[6|=1?P?te8%-=EZU ҳ iD4tBΕur/$* \rmP&AR~TD'TSGtOzkmqF\]`!,`d G xWS5*CzN8RR wE%K@)*1NꌜW켌(we 08Tqw=@x'l tأgBGs}M~VX% F8 C0yɩ,L,^ObID"0l61ښhaQZD p̐sx}r$&caظnƃm:i[nFݠJ;xkd|rt&OAQhQJn,rdN{֏5{ƔֽF8FH|;wztaE$)I^lZ=Frh cAp(hةbE*L<8DcKDpC@,s0M.ظwFKhc4XfS?~P2iO__m#9!!v.ʤ(sU˖,$,]:9e^HʺyK3@{,{^gj(^:ViCl*;TpI@Bfu⨦hao\ }fϾz/)L2SېL7USݿe7={oʖUdL̬ `ZyܕxdXoֺ{\Jsުs}Kv:ʼz8M?>SUo'ᩤްE+! HET2ofA4Vy|$i@L@ZF51D}ug=ɴϋj)BR?grlz*!E]jEƕsynF^荏\Z?>:^_fo)כe4Xo]l)BHRM f%8xr * $7 EZ0RYtc UQ*sfugdOWMFBճ,;YQ sy֌WXfav} m'iPhy4f߹!&*:t:1%Pw`\Z JderW O"'mFؗpgYlr^:!#&z*H$hӳ;%ulV;`o`$eHe8GT ͐#@#$vc_"dȕEeMfBeTNecՠY֝Ii"{M%Y"DI[KV[TT D5b4`gD!{J1HMJS`)}09OOΓ id<*ugDp=Ū}&[goT\=E;xq"w+QIID;BM?G :\ZKLԿ!EF"fDؚMz\k &aGƍvDh/-{p P_L)sJE >6]q]DV 5Ab6gz8!.lZs (a$c1wQ-DLfcI0΋a1Ƨ_]b_DžFc$.JT'HP3NhGy{aV{NI1+|g- WܢM߰B(97I8ɗ:7.0NM.{Bqu/EqxvZTܥOosvK^7?娶pIz̯'I͔=j$!xU:]{Pf gn;CAqHuZ8 inPʷOG{+m"%ܗYUZSb`BI!F%N' Dښm>.ms2EX}RpK(1IL8Ez'zY 6*"@E*E [\tϼ`ʒZ[凬]d|9smU+/Eqa\p1HjE3w9ed=+`S=#{`hZ\;s+-F) PKsxq>`)$  b>E=6!9 g.Q?a[En)15kO{{w_^^ DW~r3Wb)!ע*!AI*q8j9(zA疎 ÷Y?74z+Y>+~/=1 dX~KGR\*gyͣeq-G*hI8lڗէiOf6:nY3ЂQ5CbM60HXP"%Ŝ{*T@.KOf^Ddi˵Ҁ/O:1N'4:Adguo$#9&Mc^KwDO\Dǣ a$hra>CLʻy !rEk%,i]y>%|9m`xyS(PճwE9MUU1GEUߋypVy u,bʳhsO75 5lmGWÀE$UMSwW=07p8a6雮oҋ,>[ռ:s*E]. YF d!f0Do?@{'OP"+!~r\_p4h'T$LGD$yP?3烀\'q3[(dM5(~L\@p\Lq0f ASoo/?7OVΧaζz"=.̋e~D=Rϟχhv̚[(W5(ܲݜC@ NdBAUŠ!&;Ȗ&-8/FDqg;.k\ $ax_܌O23B\:qecϼ>꒵^^o곯;qIw|kN:vٵI+lNcM]$b/oۢYM5ܗe.I/N*mRkT]^tGqO_zu|6uCӓsPSԩ P|gu\,cUfDeڜ_D*O)JW!縜S/H?@37@WҽN~/TN0BF^_eU.Ȋb;II (d 4 nO>H! 4Ty֡f)(a S C8e7MeD.tν;cKPPkmԖhI;1}u-~"r q x`&DC!r~DH"+3Q!%NҺii˗(g^eꪴwk$Q@MrRKIHLu.Y%iNE.2vrc ' ;& AX4dP}AܾppkOdž̅H"%K8(ذ=jwʷ'F}ydž ΂6:Pp2hQa <|vHIL3hgИKmw-bY.s/BHp- O{ܱZמF5#!: R LQZ`ضض iLkS:iCX!5:#@q9@ z@k_s1C㜸 ə.]Fv7}+[i;_L|9?MZ^8—#rR3%M&g-\h ʰ&\w^eumU֫ܶJ,XeM;kSh@'j{P;KE1Q̪Z!K[ lOVs-;5 qyiˎ$Nقh[7vɓ9S#\kfB-e;GMOқBOfMx~ %;`R^xOlmXT1)׌E8;وPmr*G8 0{_p27Bm4l{VNt^,*Ne $ < "D"I =Gp!EvSOU0;?ۚ}&iP~oCNŠHժ Ӹ nxdpR_6RR>I\; B:cJ!,=Wz%Pϒ]]sr+|/W%[ Zɔ.I[S%J֐2 Q<[3̠NݧjRԦYmT  Ҧ T۹9(|rEƛ%ƨ2 @A.\% LY5%H[ ccvϓ+Qmf5.~V>rO߇H< -TxJ7n[vtr#//?ͻENsy2deBL2uB>z$1O\=3!kCd 0 2Z_{_d>?1Ij+DHLy*]颎w 8Spyfx^sMR Naj49/}150&ӓ8%.ġ}F})jٝ~F|WN|[^zac6$t & X`J0)DNri32zOw# [%w+joی](6ug)^>-׾w85i5J꬗oR3uA56|qQz9ϳTj3WVLNlN˺E 9f}>_~'6>1cUR<yPbWj_ scPdyo`˜-K}w놎үvBxpm 9v݁6]%o-C[x܏0@W$"}#jӬo ډjhIGSmS_Y_m$ f{.<^}kj.,),ftɊ?gq1d^bQaN2V}ruyuUEcq]W7^ry@zP[,Ǜ%su^d<6*a?;#ҬoB :NM_YLl{xI_5Ov6ǿ<&SL|_+[D@Twb?iv|kܲVM>2-ئAc,Ͻbs;wECM}u>msͻ<ݝnӀÉOlRu4rty@YM=sq;\N,05_v|&fw)[}5ƴDs63g{Y4}0䵉$1ԞI@kU`:tIC%h)QS%JAX(LK-MV̗'__jP{?.) ESVuUcdQXO1Mۄm@o\u'%ll"r%X~8!/`oi56%7QvYB Сe&P貟jD8<.X˶ӟߦ%-(f >'] KbYc >] 9f=ӅF)&!VtS ~|}o<:E >=au6iڽ?Տz>|;oJmېpNxt_]ZWG̬ǻ;0x3zQ۶#f=^4zbzϞ:ifbp9!]OD[̦_ve# 1ڱ7z8prz8TZhK{}I\:S5YKfd`'U i4<+Mb ;"{u 7=JNcw/;~0/x{& >8TZ #ϼ>#*!"x%K ԃSlRYC9-̌,>)_r)eՑ${p3s{<rU{[/x`$qY>SdiUϮVՅ]\$9!;;T' ̫MȎ;Լ;Լ;Լ2dH OHT#LB" RŃkJ1iIIcI(mKJ$2܃uBy!O)q-w7>ɝ|]˧I]Ip>D}u>_TSmi=ƋasyA ؏4:(+6M2'3˷BZ';miO FROA>`@RgStDJEMv9ǜ)Wr&RD6 KUQpEB2A`:T11lf#wBP>(qG-Y,Dg$zE:{dIP E.򨤨mp<Ҳ@U?Q*q N6{:'|j,95[,i)gɍce#ebNM.,Ӑ0_ŋ닳#vD1j3DQ1|,Ɓ:w& 4%궻^8WhDUMm]ȩ@pl'|(y[R."9z79/b j7vB*TLB-\)J*Yff3d&() ȒdȨkX-r9N:;"YO#cTmT Auvu:\3saoT**0 "6"knB wx Z1; $-@YNc4)HWH<(ҶRI2ـX'59P & ZU!ǎ'!ժƈ̜x>FF:Cq14E?ℋ;k 聽;|cLwr(.>&AmB28e?Saq*x'=k~ܻ"_rHF.؟7e?>Ocdw^ʛ?>-$AMya}9|X MI?RbUb]bVaW# c BRZ̆J^1=(H**l[J2&.k(ou:qۓ?r9. xt1Ĥ=4PAڒ `<ْ]89o+,Is>x?5bS|L|{M56s瀑Zԙp8 RRu (M]Mvq&ѴI4DEӔ(t%1ed$"TB }"; *&ѴI4mI^i<,ҲWέy)[*Ra `QZTh2 1k1RB*҂r6 oXe-XkPNy5uI..)v}ؕ9{:Q6:yֿٙ*wˏ;c,(jquQįhd^!4%H[ )eguK}nLPٚFlJ'JLh9Ar7)v˯ևn.S`cCRR)6{ĚUiDW,8YH@d&)B &HDdv0I_GGɼI!ICUPPk amS"5ٿ`8+X \}P }xwLyho9ZQ>jB7 am.Ym3gJU/$RwXl@Uqi؅:jHiTdDU'zT5&L1TqB$LLw 9Dk11]AxHK[` +IQj90]V.ٯN H7B'ol^v9w?: "m^"-|~<!;0T^d!RaK)SQQ6r1dUR{$v;r ;Abd;ģ #Tq!SՀ SKU,nGT0B*Caj3jX$"﵌FMFSft0T4Hm R&(PjD}a'T,ܲCuXVT7Bz U}eE؞w=tRݸ |.ڞ܁;np4٪ӵ8_e0XnٻXvpK0X`"k"QKiJĨ6xΨjgb$'|UedfKta@P.伢O E෼c3viYg)[(Šjߋ 웤`$4N~h jZ :A1㺼~kvFë[KaJuRlI̡͟|: 'EImb! u%BMFo*#XzӸ>ZcScyZLLGqsj4TaI{}cR\*͌#vE)TSS:{moNuf牘[8JM`\_<{(|*V2'ȮQ[-S!;'ji=9 kH YqyFt M 4 H9cBIV*(=3I,J&}K"S2&S( B r,[OKx0J(KׇJ7 W%lA䚲y?h.Nq~;3U +/Û8?^>VW)7Ӈ0īqA_Z|iLDnw&I0E&}?E]}8=;ޜ]ٟ.@E$䍙xB4`$'p9 H!ʐK`'X(K=-d` 윁3sv9;kw596 e2sv9;g` 윁3sv9;g` 윁smHf #o9;g` 윁3sv9; CBwT~Pk8wenkjөQ`)]Y3(>0sVc̴P-QAshHXG"\JlB Oc*~kQ@ZFgkYǧ{iӭ ze'bbDW_8vg _!nͺnlM[wHR`hD K' V4lU,u* *ƨrRgI^*m;@H"Ƒrp J9Qu9 H2ݕrL‚(,2]TBYX!,p. s }ޠ7y' ,1].$)IPCpJ2B+ \TͿN̿+/hmmyme&(4x{W]j7˗#!krK>C m 0 5J: ̙^ ^)0cG8b7(R@K L0# 1)}b2ҭ ^?<5wΤ\.*=įg~wW$ͧ}߃|1)G&GVz3"gG+xQIcwtajnm7ʿX7M?Rxa. MtELluO[S Q?z*tpb1eN;vW%FX2)OWU+S1Z|?aEb*:E9GRM}ܿ\YBC맃Ż[ ow]>.]\.qb))URTnyGş$uVb2cB/QMYB W?fo wὁ.*c=%wm6Wlͯ:low/@xĚr}M\KuC*N{txwqqjVT~6ʍeWY(ȸecFz1~?S^nHVV5zM#|=_߯)?.PrIɭT'dև[7#,PKm,|xx拓7F{p:rƌ7)- Pq .OjBd'A$Jz8ɑ Y7 b ;yĄ CUZ9}0.zmT'S0;ıppS#i>蔌¬)H/D`%fI(վ]$vQA6J!^ڠLi;I1h'mX)}꠨{ZX7tF+g(pk ,3,\`ރ+YE#^9N:X&=A}N@=38bw3tlύ1P,0[v k7luĨݗYhSl_cM9l }EW෇sн],IB(B$O 8s6xmMѰ(-q̐sx 6ڔ6ƍ|{v ݤi C/e-/"ޭW6=l1㣏bmL?X)"( 5JҦy5\:a2C=Oc,1ui-/@fŐ)(6"aਨu#Q`x+, -8> c,;U XH G3E4>pNJ7bi1gєB#+iBG78$~%޷ݭb*)J*/05|,̨_,*"Dd"Ra-t4~̜ L`xePjlt(+np4yYz4*#im;.2A1hČT (bKϚ͓ui,R#qš^rsQBKO|RU$@5 S!fG;?>cJXkdхᐌ(ڳ٧ l;'#{ %xq`Ztk7(4T[hJp9 ] Qt˂D@ X(Kj|{5vBSaKyaV9c.6ޫ3cB,S}xFMcIrʻ=▱ n 7-S RXugNx}bZ APV5%&%S_fW(pn/ꉋ1SֿSJNkP&Յt )}ܤnQ\zN:bhjWsUPA& L&nu0>#ٻ6WAo_pw@nclᐢdQĵ Xx5,TU?]UXdx=U2 Us:PsoV%RVu:>+d<,~>?%īyh1WǾ p@KpZ5jz~v2)W+fhO`HWKݟ'414=_$`n<ʞL>5/!>I*gn~ȉ/4[2~3ELe, q7<;? r2ؗ_=n xR~8,-g;7und97:ng<6*]/_c{#{#r' [1x-0{H=c"-1=c,nK9zyt1M p~10A8E*IUk8j;{~Gf<3ijf-j|zE<<~po֝k+rac-R9:AmCM s)"LJ8EX|E$(j)E^g Q2A<,c4YYMn=5?~5-q4.zI0B,XU֮M)BIU ,߫&tc&*䁅 q}mƲU@i`PnLNȣ4cuȖdPo}9Ũ1Ҕmm7U]jQvyH]mӯ- fBg}w.IQ 49F=D!1 Boݨ]hYg}Z٥=e[vN *iָcޕe5_܁{୯[?'=AA\ůSO16r*窲Vl13A=/#@T޲kY*CӳiˉjDr%ڞc<ְl6C׌ٻ^3v_Nw}@HWD`MT dIwBZ0PWwҒ6B7/bVbcϿE{û.Τ_6?NI=ެ1gBYiPN<Y)&gD.A},)M-Di)(1ɄڤVHٳa+eӉH-8-N%#X^oA CyO\q,) =wt_V鄟 ^dj6&dJt=Q7I2 3l!|Ȏ6۸'kP,K Ec!2HB dÂMV2쳩Yk R$#r,[Q,c= Z4g3 kݰگt.:ϔsݓ<ZEܪ9_ `y1FI`e!C%=ڒ5#~.9_R˒3Kd,9*zl=@ 1jP r uA*Ϋ5SgiڮAk_㞘%xAផg8 X\J^Ck%N6 ".0%cu ɤMmhDF!nNUn+qiw'̨{&X'/݋oyv~ͺ/izv8?uI/kvpN;Ybn/~zq^/e׽Eŋ9[pKW&W~نx>4/Ftlw'_T9;O/fWavWY]us'7:zW+umx!m:h|{}SU!tZ lύ/#}Vy ,%[EǢ)5 l.ALix on)"b0(]Md./j#$>YEvf&!$7X OZ+' ]X\hl c` RY>h7~I] ϮllllP{S+߰NRm.j\*Ķ( .8g{͚[qsnC[umFl$F7̠4"LRaOs`xic+y"S>o[{p'Y~-B )9IuY" qZ&ħE rBJJP DmH,e)(W11cQWpekIkiNo&fX}>!;2!| ZAet{76˻oT'o[v􎃏F h&2B3|@rfo,`g (B !)JIJ¤[CqGUS;KJ+dEFY /\JT/ 6/5ήfe9Cǟ;ص{W@elRN&4 RdB쁈?Öz#":/-%ccǾB笴nS;_΀WS1FG+R<{n Bo0>kz'|! !ą S'#.Dzc`Kc̢"$!0ʠIɶu-ҮϡHP UQF^+٤,} pFTZ6g3tdxqa wNp7>lu- >LTtwċ_ >}rȥ}O"!'+}*"D ˵,!D'xG=.4㷙ܭC6TjS1@Cs=78l.~ rcim{o9gLm< 8^:D z#܃Q m]Mw^9ho q%KO pgeJH$c| @J4Eʨ=(gd* `"VHT2 S(WDH bcL- .r*=$yno#9c_5%W{ep> ALNL88Iq9VUԇE&wi{ibugJ׏+<0Tlr.˺hG b&6ӔU9fTQS)S% RF*4ԾwJQMd2ᑚ+s@Bv2-SJIR$ glN0 E'X1 Y'i)ړ@RN=RQyX nh U3IWr`hv 89ѧSGhӣ\7܎<{Qa砣oV΁Q8C%b~K6:pIV?qb'sr U3ns7S zC7PtRPGtSjd7% ?z4 -GIKfCv -Ze Ny)/xyAA2ޠI}JQefT޵,8ÐS}S DZo2TMEHTKe9Zʥ9L26*(pō_~)/wpo(gm]hݾߴŷ| dnz.LJ*#/r M$\.pX쀪e 1a1!l 'TPuB#ET|fliĢ@خEtpړd BT %ZЦm11զulg##8FvhU C Bs򣽓s/݆8nywr?a.9suWWggʴ1!fOĄ~P*V#N\.iYn+ąٰM  ŗsH-%lZ BV|?ӞrUէ|\6\s[3-ٟC?]sey9V<Z@NɄPf QŪ \=jyUrT C GfTu!E}p/{+ ߗg݂ݻOuRFiͮ>l%ǵ5J?1lvrG?'{j'ZߞgpMFO`iO_%]w2?G܍[ol;vl})GYFo>NHR! H5*_ JvJ!9C,h(^XE%aڷ{x>i,ž]Pr9m˯sQ%@8].};~߲ǴOvk;ߠDw-?tPlAoȎj$I i@!X JVS`-2Yn?M1N(4մTQVU4YPk2A,Uy;$tGgtW;KhZ!`ӜhfQ}Tƕ1z]Fh:Yhg]lKAB5@C&SŚG;,`!ZPnm,UEzo47>O/s[pPwm!arɊYXNl2W]Pj'Mqx(6%z1mk"X*|,@0C2}TJ[GqTE/"|PfJKq) 5֥J8Re_=gUd8Δ&u=VEPLrQInHw"O5>*""r>"gecBȬ· QY"J,hS2'g[IX yu`g!-%CZ t=ADސ0m6y&A;ݿDlurWDj2 X#4ٲS6~UcA$y'Dm(Q(8]$g0U8:"&8æ_kITbL fE=a$Ï&E]K&tn9!݃.hȶ} [~K{3$vk&+0&EH8b%ʤɬdZQyY82@+I(ƚ\Ww7|vaEx Һ+~ iyTcY'Ia? %؞BjLM@_0"P١ٵD L M ]= rV&Y-GLT)0c8CTqR=c<J7_mf<ξNY}-Ae#=c5..|+zY]P5ڔM!x@,aKhvQ꼊>x[TLV%0ehľ䬍nn8_f['vNZ RYz*5=vNy4^XDnՀ,hkk7֠I%\ ((%)@k6lO%XP"d!C@^| UU:F|"ebKHP]}r6_+n5x,xG=<܀sR-: PKxKdQ5\rz@)Qbj@dGd%`75"F/Xt.ػr<>UYr1=6Kcg&8ӵVV(]aA}!-A8 sţ٘kv?O_GrU?kA;GíJ(-lv?GP||nɽ~[adeYn*!( 48P˫-M kΞ'f||fwIc !"2Qyp>ƬQ*1GkMrA p&OֵG6P{GkI'V|$jㄐJ݆8:O\\h.ks^2tG.}}=.@!)O|*2XOYe!U#V몬DCnH:lf,ڳaG*&g)z-TWo /X+})mX\Khu(92x:8^AYfM*kj=Mwnq&yIHb8{.b(l|hʫlOUH!|F 2>cl d Att:) +*$ؙ B]wB_NOd>4x⽴vg3o%\lr.˺hG ز!rdX1Siڞ12]-u_.[؆ށ1F_&2{HMw| !w;f)v}X+ ~N,l1XaAVNJ9l+b!NSC!'&n{Vǣ|J92v!g*W*Vh229m!;"NqsęO=ЦG˹=T*#nCy^04AAG(kqJ h5mt2~.)Mr U3ns7S zCc땂. цA89~<]_O{SҰCGR:4`:dˀٲ[GT$* dݧٱ;XVkF] h03 Y >U79ID&J!CL`K,TKZ#\$cӪۍ+z q\wBX}r<~|7rօ{_M^|~@6BīīNGۿ56sHb7PĄĤ`jBlgPuB UUSɲ +C(]bkO!/ Q,/kA!Tֱ͞}!ڢU 1AˏNνwt*z&FVX֞_]M).oDŽ/u1YEƖ:ܮ13qx >rq_~ߘ`36োxX8wgb\φu9"5uLb%q\lުCe,fo@xĞNǹӲ Vae:"V/܁#y %u [J:Rj֌餾14:]}]G{Lsile]3\8oӰ(LEaU գW%HN2TppmOU^"Ǧ2᷒`}imtnSm(5yji/2XE/Lwm$ISlG%z 9vq F%)u")(rѶULfE^/2#^̫l>fWql2Yޜ\lO"Jț+A%%A^eA$)K̸<^Mb~1+ S9~B|jxWA Jcgpe]ɢX<D7ʐF$bZ)*!tG BY]R&_3Ad:"E\DfPvq絋`઱]&OgJj z@O%S3FΔqMkUCK"n%>8_oTGwRAp܋cE|rGx%(9KЌU(_g\~9wpAZR0IC21)#knu {V>vMOA75Yvy!"]y?oP냞j "QSꔩp /2 <F2Uup r:\J.G(4ș P ƙdh["B31RH࠰;Ym(0ńQ94CB{@aM_KYd4Tq9ðh?>.a+v珖Uoq)qE4U!g`0 pDZVeR)L87Z{K6Bt:Q@ 蠵^tn1Rp92b9B87SZł-ʂ Yp\Z!@9*9?Z2,c,Cs/mayEsLrHɇJL\R~^ˉkk ,)X&xYL7 mzohӿ~f )xf8{9|q(^&מDJC"/'rM5c] ZBn8 p&R~:/8g^i;JG0A$K.Gv(/׉ %rc-`,=Ozzy1,wj< !d0(k .N BuRAgV/_`6wcnӬ )R'J3^%)XHڀF+Jhq!DE^ ~Փ>85p}2Fʼ7N$\y5`WBl+L#aOPxMdJ*`>B:LLQ]H9BAd kYbJ̳ KG@ӘZ+Q( ٨]Lë9^6hS~Qs-[4G~4 =?yÕU>վz~nKdu}t?(^ۛ毼ݍUTD\?l, mSğE5x1+{i#5$wU^IEb`j)q|DԜ:,n.c?+ra $}6ټ >X2Rȿ@'3ئjw9?_jPdl1왱eJ_SQh^U.[sz"֭.Ww>(5oyVtKh>ޠ-KH\+Y}w+e;)OW~]JʖVuc&.e3R`ޯLо>Iߒm?Nub<8[SD:h+~a:xxv9@-z({X.kucu_xw~~jVVdkk[eFi+,5rf}L㲅Qe/j̯naiҐ얏dk P?ՊiᏫ=\* \Jk}ϗwRve t7cȚ0{ =8<330@I)@t ֚$`B-8<0;{}p2JVϯxxrB)R&‚T{M#@@ڈTkzzͩ6'NEAzHFgg-jpAz$ =^v?|P }RE鯸"XK"J-Y"*Nv kdrzOnYd.8SD~_ v R#WA‰$gA{Z􊓎/9̠UH$M>ʋ{fr!>7,|N6Mv 『^i{^]3]aDKcD0CKLӂ{@? }Iīb~kcy{>+,ɘ¿9Cc࢕!ZkM.YHnc sn1~4ڸduզtJX<cmp  LHup8jԫb >P1@3 5!\Ia;G1ϳ 1z#4)iyBs֝!9"yuXxVJw`T%ʹ*hcG{&!3HK񼊇Iь8\:'PxlwQܜYTHVKq%OӏWaP=g6Km 0:zy셒RxFLqMhQCx wйچq֛X,{jwU ,|yboc'vszl10~t ,o4gSo:aoڻv Ѭ߁GsuuIM>3^DČ)멲QuBv Y+\͙Keyz:)gƏYXDD39r6Pa@rq͈SIwHY^6mSfU,%1;B-c>P82k$T@uJ{(L 0!i%rz.UΤ?_O{Y%mcYyut9T"D|>Wޠ{]Pvh(Xj#9I.fў-U1%'ܛ8WV l:]Ͼu+ȯʠæ v`YeIWE^BҎ_⬄D\`M%,x#Kܣx 2V:U`M>@=znwl42RP4$M' D:o*V%mNA IҠz\\OXύ$!\A2 L$Q}sNLIEx/7qM-tJXK=ku6ig7 8s5bn߾ 煠C}z>QYeSYΨB *q%ҒhKj!BM vHwGΫTr s"GΏuGA:D-F43:&RZj渏yb@I!trGSpO 2T\e 2\4aCKhD0}&D&y6Y\HyLEBncnEcJ\ciLnnM4wPTOb> 0Q}2IpBs{w~UU\A}:l젟 G(U Y0,pPӝ_>YWl|2M RhwPikgO۽QmÅ;[{Gܻ7{je؞ݺ_(xb3&,ǩou[C_gh -SjRi cu@cH?ep~1ѓX| 9u{o%ESvb?CCuwi:|~X<?/?zxJed(}qqx?.pS˗`×!߀*{' w7''"^ۙؾ>VJ_kZYi~=G\9x|ʸ,bl,o3~8vp<fP!/܈kEoNNw=8_7YoYpm/Esv{߬?(}7?\vDna秫u=Ͼy8g ^C{ak|26p7Q/_$Jvͻ^\6qmt܌w'(F7Z0u&bɇ_ -0xbdbk; /w.Vw:,\W}E+#u'E%V`x[֞'+:eDh'ښkS="DhZnEjU}Ό'.'a6*ad1XwQ^5rFEK4LEk@mYyrTH۪4݋RyS-xI+x&)T0nNk}k5x t:Uε34Iרw{u6;MtVG.ghSԻf_xSU &K3mDۤIM#bu&TgƮf`nhf :#M1E{) w-Lƣ%D~{uZxeY=ZǬJwwHZBFix9wt2@䕯1rihDթ|/d;;[_:GBF:2(j$b iu4BK{&gI{~4+C*R8鑬#D%cNօMse b|jśI$EU^܍S5ԺVJj$YۢDwIGooauطJ0g>Ch~*αRcOI=k],>YІF.\>E=uk,R rSȘgE¥:itAڲRȮPB&A եfQiDWh<-$z`rNcϨ-6ݫ N6` B`:>vNi2Qh? QLnWˮRSaIY cMpmV:E ul@ Mk -cwk.t"dst 1h CB[N+fYjXkYf  <ջVLt&J2ERilG;!o`KU * 2VjV:4 b@fPd6qsd|2zZc ΍&`V(ӉPu˒M%IcAϋuH+j+;.jPՀB+~Xd܀AS@d)Pi2KE"Y ̈́!6K$co0f|) ̚` ,QLkq+9oL:v_#*A65T4gBQ!] QQ&+x3kᑼ# 0|I6$( ^i;+Qa2uYxYQޗT n\Ϩ`[QSjkIk $DpP(46j"| $ у2 {xWZE5p4 d36^eInc3RTiĬ⤁1&bbC8I "&d߫ wnfzeRkW e]㉱AIzɖƛC G^-hJ)h+FB+Xe ,Äc*(Op(v5Ok=ȗXP p΀4D. iՠ(}ڄLkptqXGi /!!dF$gslކ`ݬ>jE#WnD|9\PHhi+IFi:ub鹋hI5йTX~:C+zgjp!GTQ~ԋB? Jtmal](xK \Xz r. fZ\"'c#͙DDl*~w=J 1Q\sw@։^ DHcTE J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%ǫ}@Fjyg@WѲKJP IY J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@h@Ė]: (\ CWJ+>F%lb%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V}J PA/w!%(`5VJQ d<+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@ݥ>\^=~8vqqfAޥZ[@.atS?NNOV?G/><:[Յ(]<]jc.C=6蒢yA1.r6uQTmo?Vsۼ+/nȫbFqR腑>_?>/_7mW_և?8o@K4_ 6 lIW?}3)%2R@gWUeKAUЬ--KqexW0ծTV߶0bﯝWa|=5gCVۂ溾 E6rk|rd0ȕo,TN}=7?u8)tl?j (挸gmՌpV5ZR^^ņn.&M_:^%? AGjv`*,8-۲\KSz.GʒC)$4*,Lv\V/{Od7ڶdPգraez ]e)oGHx  ϬJw*gsTL^.5aEl|,g܂93-"(^;Q xh DNȁoTA쏧Az[ ؂?_ǽ|z1'G\{##C8T8H !)4w( ep (@@پl3Om0H‰NrDyp|t*fuqC1g5L[ ޒ 9)N"T83rnljP@\%=Rt>s߾幾jt}լꋦmKwu?)r3RTJr`. ,,w$ݡtF]<<各QYdx(TK%,1`SbpΤłYe8 6̛OνGy$y}^ћICAD}D0"&w":ed""S\:0ʐ\r0`=t:GGa%oPtv nv* ߄7n i?nqc6(Zn#󱖑f_}! r5"d;Dng !-i }x}Z!dp*#tQ9 SKJ)9:©!x؁\Cۓ$FMםbJ252)!Fd*i0CXC":::#gG`uKtGUJ♊&?ӽ`F,5DRè s; K.ZF`jIhJ37^ᆣO٫i"HK4i9>-Ǯ]-קiZ*o=IuCfz:]`LFr!)E񐂩4o}M)U_7f Y3,ބ*Ks>aḾzB|j&;{SA JB.p•ԗ 5WרЗnn{I-4ÑwPbt#IA ܱp@dIm0kpM&Rf=USΝvzDGTdᨨƦ:}~Uv:^_*͆* a+mdUU֎U/!mLVSW#̚UZ+,^9ΡtT:Ko(otVfugfU!̡dpnޕoٶC %x8R_vQϻ<(Nj.fe^o9f讧뫅2r l1a;mOX L=7h`-ZY32 , @;ovvq,Jck\Y㧇=|@}\@{t ҟAѡ&p02D~T=е* ~{X-e`-yM>jb wtJĨoFV;||^||R&6@{4aȄArl/=&ua\8g_u`x sZD)3ln%9 ke=ǔ٧{߻gݟ]p\bUqƸBXjJ#UݐJx P<4䡹) pmcv; ?vG]]+~`@bVP?tz7)c3*sX;pAXce|1eLgG6e>12f=RSYILɱ-u[Gđ1reLt 麸ܾCH98>ro;/{2h*Lmґ2ogn2m0 w|$K-of ,, F!- YgWMx^Y)Y" #_eaVCe8^go7d 5>{F]Ͼne9Oª4LyzQ6UTAKtixot3Ja:z1Mt(.s';ǹW(x9!Abh4xf ˂6,Y@e/f p]bUsH쒂L Աu##PL:xI!֧Lb([$-h|CAުM{[g˱#2\*밊a,ƲHAn#<#EDM7{,V:,*CMgVM*@aAK2A) {8 "f{D2v@2v 1kJ5h+!n9F:$>Aɇ9H/.ѱ!1QQ9εL!#Ҧc DL'9^9 5e8qfx *8hѦ$2""ԙ8nht1Ӧ</O7LRf Ν6ֈJFrca{{ŸB22BҭAEV6M;HG*#R"%1 s2JaQiƝV2GI$k<pDQ?&_]_ؠ^˔8D@JG]$ 'B=7Xb@"2HIxˡm3Pd&8FPm$3Zk@:LuGAVq@pcH4,02G&4;',LclRg-ޒ l^Ay@1~YHI #e{)1ggUqӕ*mj]=;}^VyQQ{"U1wh=9>YG!Qf:KkG?MA ^UQjgVIf;b֩'(_C1#sP|\;20 Z_}̆m^9yQ0+*<6k)b(zLOC֝(AEShɟ)PIAwm6:@DŬOEX]k6QgH=Z.>/n竔D-4\ԛ꿾>ze9 _)C܃e|/XJJծrsʺw}'$uc2ϋj܁ ;+<0Bd'4|u_ry{ѺzH|G]vbkntض28-,IgYn.mvfYZjJԌR)ٛ P$k]X*3ئTPPY7oahًqU;m5.x2/k(FjŜ_(eGͶAe[GM`r\'iM ,k;{aCRٟ079;L뤎1 gJ .Tɓrj#bcoZI~'~Hpc/(ީNGLe!ֳE|2LC38/)՗4iV]I#Ez4N7\j].raS6J!۠LnŃ3:'N3:Gd6Z8#FُK.pyW#DGr+)Nz-gz9_)%}4`,vf^ #5դ}#Ql[b%2#Ȋp51:U w0X?3nWZd?;S~_i8zG/Iᙨݯ-uIvѕfZ2 Dz 9C[Z_r) eco>HWs#cP/q E#6249NJ$5Z淟yr]nQ:'XzTv_$ݼ[{G~) 1ؙ~L1 % Ғv̸Wm2<3ƂI>UT:+$PtaF`|E!AAqt2&,"%C&j1F p$!VnJrVe+<sKJ& "k ƒ,)<ڤ9` -xN`;Kn)} O>>xuxovPuoKTʊ7Pw[g-tI~ x1'XN)M%y ֨fZXMHvG1xl|G =#|mC.=; -*91%~i~9bGD^@A>p>z⁏#*}ɯۛ>{^~`-7n6<0A%|[殇!U0x B [kw`ue%Xi؏:fVHi#{kUaӤͧ&,?.Ц%GrqbΠQ;$6vNͦ=}os'])|npX4q2]W2@CM h_=Q|=^j]e#:|qpî'NMm"D}q}aV1[){aI2MLBFTZm*0譌2'1qkřw/RCDoxΛ?:x7ނ LI>DɈE2QFGm^8RHcG+=+jq փz&EE\Y6p Ń@\|s0Z~֯%T [ [oEELJD n$xDғ*1HĞCPNP `A`8fJAFUiDqZh齳VaIKW&x8Hm>ynԼ fIY42tE6 r.ۦOLg[<">ո [!IhIu_t/ʎ1:g\ثa6XGW\=\6dÞ]:si;ZTX0h9* xa1[NkoGoL>GR˄\V1Bn"SP2# '?a\8HuJ0̋ˮxtt)wbCvvUʚD>QHAa :Q*F"y-2XقN>ɓ*Ub<Hx]uz@p[+NF2*L,TjGUyZ_b W[b.W E:i?[dYS!D @Cn}mp>YuWX1W1ȓpƹɉԬ2Z8l&)0&l >1/<6jU0VrER}jw SkDPL$KA'!nwU\)F+|yu jh,Գ.S\`,1΄`2C-BΥBpp Y^ U,0 TKa< 04^83iR=:Kyɂ8YgW}f:}CQ g ŜshFi9 Af eMmH|9Y 4w?8:9;Hָx#`^ gMVbf`PFaM]jVZ^q'_ABeV9eMiƘ+LxӲLho+8Y._벵 >(R)Q0% R[PX-Fh8N& fgf-IFk#D8㫦"l|T̲Z /I+CoϠ{1' U| IH(a2<kXG5 9:TeP&,)\_n7jvM_A>JHUB˜l+D.DhO2L}paտ׹B& yZd>jd<#q:2Ώ$9ɑ#edce%eas=dg,%&g93d$ʺ[/  ]=@˘S$RE.1@{LV;!lN ʤo7mtoZyt"S\|agc蕇`ِV U {! @H@2xHZ::Gkägj5qK"CR ?UpZMO%gBf;]?&۠d9} N.[EGSGXO5Qy]'}م*0S61jк-FI KUEF,#+lƹ 2%) ;D,dÍ MTXMXW4X,4PXx[RIoy_uj\|\zzX~ 0 C49 urs:rR12oՉEnv$*QdBdT4*rF)cΣ Iy %*#v5q#vMFy6okiǹQ{`w&Gdp rB6(qCEf)f9()=Dl]L@_/͟ɶ>E̛aiXz_?l5O1(@JzUA*gj;ʫXaVk,55LjZCM4C >og9|^alITP'm#D(b)Q/ PWl?d/ҳtQ7OPޛJF둵RتMCRI(eJgZx8_An~Ttnq^g/:O+T8eg8|H$EICA G=ꪮ2jg Õ':+5hy8)-;O8^3v$7$nRؔ;8WRM.0˔eۻTjAH! P3P<0),8,Yֳ;;ǝ('/5T2}\Sbl}[VA'q3BVBGц ʀt9d E]ϼ0+B:M}Mhb{  #ePbqU?3y2tn aKYwLrhGTOJ_ 38(^)ALѕ"GH8}=ֳg$J!#+'HX q%#QQ&C*ЗQٯE烺}{ŕh햲Z̗+>ȅ*3RG ,5.ZseI)'!N^W9u} E:lu-1 uw "[Yju;]_vU̽i6XYmNCWфDu/klTϓ,2UCK§iYN@Kc,s%7H)6xhK|–_lsO2{`G>l^z4^h!dľ T} ~=h;v~)qUYMh/1|{!l\ʏ q5HxsO{U?k;7^M?ǯ1͡S%[k G3 qvX8 CreCA4L o:Iv|sR'cDzc՞^ TK`^#LrCL pR0$EpLG8|RD:6F TɝFwV[8ńQ(4]P`H:K@go:T+'jO+ ǘ%`@ DZ`^n}iMQ@t[BSSPX5!R:(M\XZ{UpIw][f:UZeE 9"xߤ5Nw50J-ǗT:96a5849aVNw^_NlpM"JJȕۘC*6z@s84X7̸qO4a8m,!:"F|z7x_CJclgpfXa a5w Cאd8hGQO{Pa L;ȁlu^\DIJ{!ZzUc05:: ||utNwZ[yx-X$Ww=Y[v] 7#ffs׹s6>eme[غn]67w9tfm^ͺClYb˶nZ;=4 |Znxͮ/{:޷Sl7'nzW|Fۦz\rgM&)R'm/tIO>ya򶣓8do]8r8a^:n6?> \Z ?MwxK(?zf8֎axrW|uFO櫢8E\3IM!A'T8ᎌyQǛgMj#3R:(~O"NpzXԼ_XԹ1>k1Y`2hhb7ef|ꈅQ sR%6[#f=04Ȏ}R*Fa'T$TLتnB{ M9Ʉ!G N> w::r@bHfFH17iL4Fӌuͣdbl8?q״Dh+.iGsb.p, ϕR%Ss6Msjǿ?w64KQ NѬע:&{1*@8R_k倞ݜQˏhAMkaecx'H"R_d32ԺLT5mYz.//]6[V2Vq#o{YJw4p!&1 =7'1Ah*DcYB$f(hc=Dw>%9:@Р<~ ZGh}|*.s]>1%NSPi(?Yw'9^X-XT')Uhc9Te~X iM`4|2~|ٍu FiT)[J|9I"b)Pylxlϓ">)Wm^h+{7zG 9ܙ#qvpgGnCf}$ҐK@n[P\m[FJ|n>j,߆p+ƧkeG{JAfE ! O\,@ ^;hp9l 4Q KA 'ms}2)%x,$m@xU6"!J.2(kP~l?&4bS \Y!Q!rd 'pp0 GtC:\I 4\4ׅޢ9RRQB4 q[bu† +뫳|kzt4{5ˑh%mo#(y<,RgIwDy'kguhTyw2wIfB;b6XFbQ)qQsCDst|fii2[_Om=Y8\4%׍Ft~*m42VT1{+tr))J2u_P9q9>1x߉>D"I?G-{)/ӷ4~O:)Uu'ǷnrQTe?Q8J֧˯{3{6s~Jfh|s|fD5ll3 uaFqw>阬aj$2Z|Ȧ[TmZ [ﲼ. 6^]VtY|-R묃4zGG'(fe HVj3ɦ]ܳ6XlaJٓ,`9pl%6ڊ.D2/K(+<9}#5*(jj>Nb}|63- fYJ=ރ&=[ >,_T)@t ֚$`B-8<9$ad0jdYsxx'RX'Lɩ^GA=F?R'Ks֧`l?Uū})M[ G$ExJOz  A22S?A)W2SʊJc.(]Զtgp}vZ(~VYxRiu3> ιF W ]2T! Uq"Y⤝wzl|H;37$mS>s2ŃͩP>wIHe$Ԍ^l۽tF-?;x?J-mZ;ÏE %:zkw.B M̄6bq&pPui1L6XR b 딩DAǒl2k)(bLl8=5u iK$,J:c!=S+kbK%ie]Ce̪ch&Kh1y;"9;UvH>_Ɨ{j37`Cog~r˫m^(kD^)mJKk X 9+<9+tH[:$(qxRNlcHJ׀˜=DtY{c} -lP8T0OEU/B6K-x5I^lw𦭬րh{E!Y/H!:Y%O ƿߍl_Ń y^shuu]Yg^6+VbO$I+)0<$lTۉ,6)6X\_Cgt+?k//r;-.n.?ŪC}Ce=cezP=Iu1m8Dl8X{hϕPg^},}Y?,#ϕHx(sI5=x8/cn`b0pch l6T֟]_a#_& ٦J9TK۽Ty\&bFyw߆Nʱ'~AOاN ::ܩng?~+tml߆ wHquX\Sߑ,0ʯ_͎gfuy6tW'91bV .O,†l>vH!XW=||t˫ۍ7xaoN:vJHLSD0([,2IBRG| Ŝ,;X6쌋X[p[7a.}MqCw/;Ӝ8-: Po}lPڢ*lx"N5at!!KB;J͟lw8כtU"z VRtj(Z dQYH(|DEa--T(La 0-2rcz>B[0u4"Vjե&6?T)LF# R$˺„dM pTtov7ݵ<'F+hۓ. K6K#F:B̨B%"LrbTgƨ:JQIdbE‚6$9ZZFKHYH&BJW+nssѰj!y!OASAJ $i ;Ih&~@W}iy~"|KUA!,]>ʯfyKnQC/|xwCui#(- bfI/#5jxR|ϓ6ܱjg}" )(^v@Rjk.xQ{R,a*VM, LAHVd\RA3 ʣsAh"#:& A7fhg]l[xd*ѯ`KRJ S<( x]&/MEC-Ck&'od-MvLo_G}~cS&{IAz ޳a<ŰlD' E҉P9' jē&tg.L:'x2uNI! 9ei39J`ABG1f:oA sf%( DTFc1j)q=O7ˏ?GoDDgd0P([@*X2"~IZ  S)m m2%9'ռPs-WvK .6ӓc,^/\ɦa۝k:y|po J}V:|{GIbo/Šd)UHQ =,$d Y(-hL֒LQj_7;bTzԲ7SR%bd a(e]De'36gglUfq/}|{u\`Եqu7%!8-̮V7c#Ԣ LL( w쁈a Z'd Eݶސ*c+tJJ6uBP\͑'vR XH";ߺP+q{촘s4;c,^N^{B;ji z#c,)n*":kZd4#XEL:u1'!32,:ktk l*k ( B̤]lL'>1#6ӏc=k'3)6ok,mF9$s44eTEW<()kmme0h@@I](~&= (F33i OB}{f_.:_\wgSZg3-9/~O~q;k 聭;xrXtQL$ 5)#4kOGJ;q.^Ň ~ȃwF #x(-MяoVd v_I xA:/VR(?a0e |IZMk7y0ɏaS+RtNZBJ Z W (JkH\ 8jE%B 9DI$4YE1nLБ}Ⲯ1xہnz`3G5w3|?:R5"3.ZBMYWH]j[\jZEl/dMwI3TaOٓ59>Z/%=H*;S8{>krglqؽJatT@y+nqS[EW(3dd{~FhD IOI;äOPBT;ejI1EDwpYdz;?_h`mCi H)(mRb 1kS (7D;5i`Tz?`_kuu3ס#5ݩ ǜ'񖉻wd{s*^3ǔ*6eL-+,gԀbSH.8]^EAA4}0 #Iqb8Cw%~Z%PEU _CjhQ1`[4{JzR'Է5;e*^ _cRNf²!RB7ɇi֥MA_lw7 \ja>,<ޜR.u]:nW ~2p"UƭU?(*R_"ߊMg\;as <(<TjU'mnm8ۡ*$dUGEA|m977 8z.G1ۄ!=D9#h1',)\nfvFMg<$t^.TPFfAɆ4I~5qE:Q|(6+g2 Ek 6Em)q\t|Pf@@#)6gF\(lw1(J:k#Dn?B7} & z1"ԋG{˧jf2R2Wf`A.^^\J װ՘ނ;{I??K+G@ɾ\p46Oǘ9F Yi؈߽:?*|nʻ_Η8qd-"/Q5UQeM&Xs45Arh?pָQ`To&j%9<,oA_M8_7x ND;9(|;VnQ/ɋ$&?ߖ7pnh2&e=z40_mxN eZ@yq *sCmy7Lз*aXB T|"' ļ~3޴gЦ/[?[xES$Di3ПayxsK2X~T2T,[ᡘ01G&}eɻ^Ee F).v$ rI@EPB'{+pO"0e&G e/{ T/y0y?d}+$`^]܎7:̎5k[5f|ڷJ5l/ /^0+Y Z] 4Y4jDgƳ±]R7o~9tNmt)-:Y&Q W'x4%z=+xw36#*p8ɝjs{fOt yzhu8켒}<89+S]uf2m:Vx+,|:EYm8QfI-*r 0k fƢd\ڵ.jN[E?&RTsj"A4QSK-ag?L4%`X~YLSn~_g9ʿi:g=v$V}H'J0O=yx qzQ՞p6-}sc>h]dEuN .a;cg $;vztlmwbS";Rڅ?˛7SLfxQ{0[Hqr ;$HKRMTHkF[!d pJo:(Ĝrϥ`hINu%9zp{9&0$^#5 By yb *,ku&C0!B*Caj3jX$"﵌FMFS0OHKu6rKB\plg*q6Ed&r*X{[9RW{p3+UZWnзγK`AZ@=G(i1vSROr*k'eSzil|k,7l 2zHTZ_ jeiy/3vX^*._LTQ >(CJķهPRlY> 3|-ѕ/0~tr/!x:t ^k2k(~ P*Nt^fMbҾl-Kze5zA.i 7H|dYg6_Ŷgsh]?Ϋ֛[:\,o7f!̡e^skn7-獷w^iB2?otQ|Mx6f'mꖎ*/JYo8tӦMOWw m6_e ZoIMu ]E ֤0E: .#9\e<,E=L>*c7%@ #HE歠@PT]QmGs+:2,"rئc"V_"DtAZvu:&%Zkqy9jj!@xvCcG!j%A'>Dt(;<]yKH3*s8gܠ`r2>YĖMF(tZL R3- 6(I"#NHHyA]j1k Pޤk}HRm0RSJx}J!,E NnۇpZRǽ>0aD8U1`12ogn2m0 RG8o_Zf*+Y@UYnA G!),wEI!ٯ՟]Qd_'f0M?H9F/_NkYq(A8)oMz-xɽN^'i$''p\.SA|4,~uno^~\qaݶwg`J5*yo- j# ˿&JpH{qdg ]>$!si,3|^eg'Efg7:Ȇ(2ts镽LqBЛT'CTE/ `8?a2+e?3~QYMADgo2(lS\g% (/%y6[#l=U2G{ݿ#ZoTHe=`sKTPQ3E]!q :M'\#o*aSgH26뻏31s"X϶ =|]#adyoįa~ggdj-~5٥;كER)b|UB˯޵6r$B/_Y!`싁rE?$bLJZ;{DJ5XӜgwկKwh?||Baտ槗5v-׺ao%3 %C~Ѣp(}Pu*tVq$`A[wb# 2sY PgnCN:,*Cmefil!hJƍȜ' Yi`EL2v@2viEP *g# tH1s_8ZcCR YJC7ݝC̥'q86 z*^cC/:xLZ] X' 1ʌ\:^2A}(%'2g-t%hQ$Q4a<% ""I:{t0Ŏ-f?'8}n6$F_N‘yh>:Ä$7jcZjz|nUj3/jjXW0]WVQ|EΚސg%:59$IQ3(@ˬ\$K(0uo=6Ce/Ż8ԵoHr@25$d݋2-NTr:u|-PHY#Jv]&@-$bN6㱖;y#}TEǧI?㳷[D;k.~n}tŗ>%^v &FW% fƭYSگ?ۥ;3~zW$*. 2ۛh֎Me>6/0_>/o.X_$ïo٦?tz25lE[*ۭ߲[u}~Ϋln,bvVhz*[Rwؠ-~M'?5C+ƭmUl.Y(ҲZk:g@8w$ɼܠg(v^OhJl?_RRK=g(Pp> SO{&ڞ ۹݀|^M޿(~HRHnՐ<)H v>UZ\`9I JasRLW\QgiBtO褟I?S1Fi%o p@LUHpDvIi{s'bvQF(}6deN*3-y)!س9w4X ܻZL]?'`Zȝ |gw8yG]H %eg(K0"n)ԪڀHkԄ&><OK::/d޵7LG~mz٤>5@)]e7_wk?Mƹ݀p.h"zu.כ>ѯ0,~~44-"O|I\uk/\c,#N(R5)1]R8 3.bbGLgyP=[UϠy^!yXG  4,c!AyY!٤8Ys|RIrupL%dj'1K4ZCa%HFb2(d NRx( $[dvWǡ8Cwu/n&"{PEA7oA4}I˾y0R Ѱ? -ȅn\ 5~5Oo봖wGǚm?~Z t@:9J|'dMϏJ~Ij~̏0췺Pk57NpdZ1n,1$4+滚}lyt28鼷I{R47$>#fL|e?MV:̻]6[yo&V<e&)gy7Z%jrod9wcy1KFۄ!؇ѻi.P$^ٺ'9-ӬJ5A1Jzi䳒cR?lh^b˃\.QA޷ T#Ϸ,Оw(%n/3br.,z^y7RwuZn"LTP@jOe ]۝[CeۍC^ֻ/?nvjl)H}R] vMǛA~Kf5/jB !/M%9-Lԓۿ7Ԏ.'kt͖`sմBQ|V ߝMws}ʹc#v#&*']Ԏw܌݄c1MܼE o< ַ7"jݫi\g9K O3h80y!KTUlJ&CO(s28\z:X{y9B=p aCp^@2[))gHebLхX{K! (Q6TeoesP6,2/=.ar:*Ii_O椇iR*+ZĤDQl[wVڧp?:Ǯ `Ag8fJAFUiDq3nYyy߬i7yE),R_gF4r +& ل f4Z6_7ۖYM>M_&[ EϘ Ps*+?)ݧ/rv=JҲ-Jt;fxwƃ|bUFYaFZcqdTZ/O0@"puHoAO?3'sG$k\VΑ`0h&)I 5AXW7W+] q']{eP&oUu|r22>yOqZt&80qP(:.'^H.S&SaJ>e%"mY,8N& fZ ߎv9jBEzc}O,|]HӶ>l(K 3hỉBB&ģ?X8cUv(+gPqr^ k gGqT G|%$ &ϔlH{3/QWrA<\"yu,µA0G^)r: O"x5vnjV[_&ɣtݓ;b lںѳ,l.BCs=gF[D֋}\t䨸9HD*\ [3DhaBnoX*p>@`ﵕ3t٧3&bljW(_ӄz!9E6ĕ9Hy YDnH?{׍dc>0,;Ll~ E2X-9ݭ德uU]g;~MZmT$y) k^T@fd,g%NcgrwKP;z\.||s,\Žn~8 ^8_n*Dv\t$}ڴ3SICj[ "֧j4C,3δ ) 4q}*JnV ^8͜8LiX, XɏE<` ۋM]>rooX釓Pgi$UvJ):NH2hrL26Oa=h_Kq !jƱGCۙon4Ɉ=͜]Nk{Ҽ %$9t֝uH @6J6%!BP`շhp%KoQ@J\BT(7c1NFx@"޻)w5Hkh\ XN-Bmj(TԻ"A7&os3+[ѷ %(",8v2"N3g?"~ ps%<xt-#gDf靹CrS:(B7p?g0qw<<̏Oa{*mmrvŵ5ZIT>қ_곉6Ĝq} Ro-9Ѓ^ٹZuʀ1 4>zͼ҃5.rǫ|v^ʟs)W) (:P@([bR4j2Q'r)co_CBfja[>_?4}o!,z9A|!}Q똟OO mm[E@+'x}^ *VVk bW..W#𥳐55BPZt lspe.o]Pks[O ,ߥc/e&?MÍ~!c2׷g[:?A#^<@U˘Vµĭ^y3ʈu~zϔ+wNh|VFVALl\ e tfHT3[tF%S.ǜ;Щ"VgdwDb}4sSU B%KT:[fgyÁg~ "7ǁb~؄xBz|夝c4?yh<󿚞0v7d/|{|79Mq9]|~z?E`a3~ݞfs}/O6OkӬJsƲ`<,O9*M8\owè^ڑ/|#[J ]&[w;mYt1kWHylء{:~l訋G۝۳-oKޯeHGvI˃^xԱ>ڑ{19(!"Iھ͑ W詻V㎬'nPT++oͲq.p|rOg>a "C2!8J=x8i髰P2oJ{HbZK)mfvw&9\#ݲb*\?Ѐ/?X5Q~0$.Vɶ^"hJ߅d9;CrNܵzS-1"ljcHMhE7R rcQJNwcTz45>w6-5elݍvi-#\0UM|Ξ~#7ߎ 4ZDnNWղ/`RcRfI1=g$\ݥ>j`3;,E]l@r"bBqCO0aG[83erշ:.sqƌ ܣcy=n#MGYnNDW6)&Q=޴k` %sǟ19IJQZḹhQ6V:WEGvf 3 DV>'Cl8c5E~ Z},2r6xrGJ-[ +͓Y% bi! 9DtHW#KfHKv1LcM( TCJ0@LG9"xHV*08﹅h](5l*-xyQOdjdmHMDΧ`tkVk"Yb^dEc `jZcC qȎ.BXK-#og L#D˂`4QAV261L@ʙR,-ϰEg{Ns xc0ZCG.8;ǩnDI`yU%єxX] x :p2m,}n:@C1*)g*N%ԞдUAu1'G9!.⋦ ]!AE@HE!2*!:EA>DO8j 32]^6"h.HPD4_qʚxnGUP>Vۀ8Ɔ3.NeS Yߙ}ɵ _Eܱi"pn۠&*A5%8Tx.mwVՁ7wcI6#SQ o xhD.8z1U@@"Q9@wAKb.H߹C׀R"e` pN\@n5G(Z.^Ř+y AnAz>^AZvAR橜 c|(:Cw&H"cP<7cecm\g*HT!kԌ4 (= V &@Q`WZuc<\JR*1`2 @mIlmVNgX;FH, N#&B5we2cQ7Q-jpd`wJX6y/^4Y* 3cbk}.17 e"GaD#qg%XG:8 o7-F #͢Ьn̵JusHY]"8Z Ș :rop|۩4cFc!r*^z m %t 9YGM{0P}s<8(*'+I$XJε[NnE!GHSySTx]m@I3 YFz*:jc@`r֋X:X qjcqϦ%_19}Q>\9xbhHƚ r&Ɏ#Xo 9vesǠRݥƐXo\TsѪf(m@ ZxXBՄgܚ6WXxJNiU!+)fcWlYTJ(RKP[Oπ#:ܶ{f00L˂Z ^ z0W&="i >%Lܰ0#Н1>D`8>iez7VKczRM$8Ui)v;ἅ>`uLi7)YepD$)!<=uAşzQ˳F 4 Vikt,\U`!މ( 3BjQ 0pazU.n:x[;Ib͋X@?ĩ`2fKfTнwFYb2ݳ8M)49Qq<_d<,0xdbS]?s6~?9Om5 Pڂ)p+V/)G>Wp +Uo+h`)2QW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!puP#{u4W\ s pW(4W\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!p!+d+`m{\Z ]BDׁ+J*+!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\ >W =՝ p5WtW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!pW\!p&fwcCd wVyU zsYKֆ.Yu?\8͠QN>z뺛8RݏE3:H4Z͎ ʇ dUNjײ"W,{a4+bGc"MF^rq#,k)-XIq;b/ 5Ϥl2&b! q:;62u ņ`:/]j$OͷpfxwE|"糿PÕhE5M3Mȁ|GZ^rL e |l:˞\'="Ds¸IΟbP<~M{lX|pNg:4 4^)ܗ0K/n'O{5z:9V7ܫi_s){}~mibuV6 Tl6ƭ)y1H,qƍUsDOwEtm>=~t0~*`IazrۨK/};Mp1ޙVw[= Ї]S}ˢQ2m}l Koxi9 ny4՟F/1?,X!#k_D.3'Ҵtٟ⏏{KaR/rYWrMrTd^8e~AHSTһ{^@ޠNna>%:v S4VLg{vb'weY9(ۯ(C^\És,*8~ModHz2\HMܲi-XˬiAiKjJfڿLu\ Fq(h(jZB4A_@KjejЮ :czMnaj +謋EvӒΖ:?f/1^ڇo}Bwɧ$ɫw w+Xҁ;FE|̮IMX0ib:딙DKukP/vЅj}^Vۇe_ u\-/i_h}Iwwo~Hy:Wނi,Kmjy; nkQ#cUd8rU#0~`hu aD*hƞXEFWBQS1t{o^roO sAPyϣV}?|sxLRP󌪒ek_it.XQr)8TdI$1TM3AбƊ۵iJ SP% S)U—Z$Y) K-Dp^$YjTRd]Mϧ6$U97H0 ?jy 0ҟ0mޝ2Sf1YAZMvef%&qҢiTyUtd0 -~4L0;흶<N{;mlSR@o 2y"9S}٪ ^n&H{%S則>TNA(e2XZ0#pg7cf;5C1H˜jr=*ս] 3 +~ky|%Wd8rih<~Fta^Ь!KfJb:aaJc<eKP= Z}hHU :|(;JWe\QXL, 4U$(u!$EJ* QeRV}Q)S&j~|R_pGkg [g׃iR]$ͫtXF 'e9[|E;ݼ7+f|oTrS4e"Tq=,tv>j{+W+֕hRn`ג`uVݗarQޛ  eZ{Y 2SA@o *{ RY}db/>k\ϖ&3Dj T>:AfjH)0pC˒0tx Z{hPvH0Fs-Zpȁ&  BLV4A0;5/!fj);,p~H{%SZt^Y%E&WܡD?dZojit!ËǠx}R$ye0ҋ2 yuV?%/?zwG.Cݑ-mD502 _BWT9L7.4}- .Q,aXI!(qRy-IEYB+*H҄/R|&yqJ@q:WLSkP[f-a}}zKG1Mz0eB,-eZBu5&IĤoTY*f;FeVǵznVۧ1[k6fV !`[.ߴr2j=ڠ,\>LqwGg.gSfͧRWu-̯)/ÝjqS*o4|qEnY [pvɷwpa*&<i5x j8fͯG#ޥr%ldޛݏݩsmWVZFnOIMX;U;Ε_{2$FM%*zDvkĤ>PAƫNyvEИ 7N[y4XڛJA&=lۛ%yLxKΓD_;mUp3tȵJ=)OS2:L*N{yQLaijJm)"+smziP2(WBrt!6O#7ApI1p%eO,خ:fR'lWR o\2)P'u+)K$kZK d)lnʝF[?b0d4L߽>OnV3p}.輥 % 7wۣlYl_jյەa°ްe/wzYR-n"q?B\}Oݫ9 .6$X>k˒)ܗLa?9oA5!;- Hr t Y/,LVgdUS{~,5~eX\5Bj w[ts{O8$iwV>FW?,X`ve{))3eQC(˶dSHYAu#`qӮ:"#/XKź2x5zB7,þN˴0UsQ= (i^@ /bkiY /r-MSK2Ԫ'^Jɳ,,c:yU"gXʼ:5(0&$d7:R'חB/~Aݐ!\9I Wa+T: Qw-u7ʟ!)r}вS ?6cUŻVCcZe_txd$E2SNl>o]?{^Ifwy1MCp/ϫrIC6:< +N^@;Xy o[d~tKUrUU\10.\`fZ}tF|YڵbB-.ATDb=A~v\pS B_?jHq8- 08Q[K8dl[~z_6+j$2oO؇1B뾝(??V,USx) !=1' V&3P!ɞءYC}@Z\>O|!|E 10+{` j* @(ucK/@/G^pKRnחVN w}Ci[m4=^ ^l ~L/łLk@rF#vN h\϶W[v ?SӦKXmNq' ֜q#R^ų:Ʒ~o \C֩iZBvv"Rz=QQ;~Mu2',cs!lϩ%"E.D^>OTk8z{z}=hBZ32#nťh`q|̘Lan i,yaP OL \|Ƈ7g{'QpUGAd_Q\Jj)JGMRqh|WiL:p+bm]X ^'/0JRv780)5߮rh}5+ԂWҋP*qJ{Ϫ7uŨZW ҙ}R7J4lr +#HDf>m9oȉ,Լ sDB+6F8!o(J?$Hr(%kIL,j !3W -v>/גM:`8%1ͣ^MOp9V#*=%p~sv:}I:ª9S* t1[+x.;~:u'Bx70J2J f]u肁F8"k:*U^`Z:F: bqf˾ AߟE}5bI'FUJ)c)?\(ciQnj.pK\=M;*K>+w9JZxD<FYܴ},"p=18fe'h[}|E6ňkOf\ksdoM~vټ06{@--ca`՜`U{H%tlBLHy\E͕kRlBA@a ev">QL6:&=>VuH_Fͼ r+%aYyxQ#*gԮ_{yMIKAJ jn^qe`MA:z-s?$ٯ!e(=pH<&. A&3| l29/N̒Q@dLLr5@(eXׯi>F.5uxnDJL{T'Ql H "Sk$%0_SeGi$S9(D{x u~"#$ r*-Per +6eC$hs֍JXţr$;(:3"Tpq"Ҝ(#1뗮4F)E)=w(\vRYEW@ML,cj4sYf;u)("k .dn2f62e eD!k$!`k`T+4c^R7 mDuU?*׽Mlu Lu H"k|nb7~Kҭ4d<*I% &CV\YUSm&-f@5ʹ,y؆i.r7Bn5m' Ӿv)bc{LN8ܵ~FI~4LRgY@KĘJM4l7)T΍T9 \i^0dѮg}]p)2xb|Od\^Ʈs2S1-MFs1_I[,:B (2T!h.\Iji-u(͙R`8|5¹ވ[WH'YCy19djhxM0S2Ү #UpSQҊA+.U V&F#;O@:(믜{oS/~Y( {qӜ&//SޯԺI-39N4!%IeLޠwB<&(,4RX9e/ ܀QKRSJNe#K/ /[mY3Y%t'4uH/>&I*ZhA(C$5hϱ?5zצ9td9%5f򤂱xJK$42q6{N1{$XeSE,aPEUdl6>AhtrvvI 4}nM-|~,!Nq5<6yB_c >SHw7f0fZ%Sv2iAzqfQf)Z81[r&JyeOYxm vN1V!|)Nvn͘wHvr*N$nWb][^D0 lLMG3*C6i!/^N6@n0P\%"ձZi 0R,CK,#H ΫR" BDILBX)MF$ƛ̚(AJGhht_A^Y͔GjN$lYoғ3ϦE^o#/au|% Su).!6> "MY-u\ԡl'>.}rIb6jV|̩-Gݨ.ϑ۳ݔt60JD& 3f_AS3 R%Ҕny ZSvY}HCޏ+ ^Ĥ=!9R^]Ԑ5nSzi(;-kܔ3v]fa _3r wnr iYƲC n]rzNM0^F4qs{:SPC*f?hr9~#3fr;4sJjY%s1wR&|oޘ>si|p<*p.^S^:%H:5+r!iU(򄐥ɲq+w_q5]%K=$l<+W[H"*_LZ빕!<+)So|i\=&|$^.bޔ~/bt89v∨SQmnK;ܚ_C>Z^; J] iX>BU2ʥJ}Ngw~n>})LKey,sM5ύyfem2GVT vsc2ԓ|{Zraײ-}ZRF @L$ŖWȚdr3=_,50w${N}_1 PQZJ1V _Mᡠ !ensEUSɰ$/*2f^Xf1> cQ$Xx V/a?a֤>ح8K)cinA:Kdr>!{ѲFu-Gu9LKҤ}??zLo|~tʤLAT tKk oʏD\SԿV c*Ր^ xV(!|P44!#XѸFmeJiJx/?:Z p#s&U~U#/)ٯ P6Ie\4fMo~찝Vq8$~ntgP`=x_QwVAH&i>4dӮN0+L0ӞKEpªk%r (d?>gQH5фzwmhv cG`1n̗=LJEcQ,[U/)v(U$KrN:"Y=z3qXM@=TWv4y7ygH\t-qMq,*>kTTQqΝU)V/JHjCr3C ./k̚J_WIMŝΓ描+dӷr+dP|A2?hB LadJa JKbwN%"8EIPų=}ے7 C79 ̋cP݁ȞQ@DXq$ rt6c.p]AI &>CjI$=&p(D,T3`뻙Y}ޒ[w3ȩPj{55t0J"IԷH–tϟq%~ *\\gBm]_ci|ي]SuخSl+h -}z9gs*[Ƥ0޹[8M]Sf$[n1A(ToZ *ڷUk)e!&.W3>#H³DX(T&OޕPęnqABeK̦ 76 X; L]>KH+iаQ*q%0&Q&e7j4(Ac3^rHXEidHoVj/ce4R^VǑ![/ }BRq.3TsMϏMsR|Vʍ߁a6 fPDhJ}-[s>(Ҿ3іs|2 *꨼;]m'٢Zb^5B=Dd}YKE\W=ۉ8lde8yeY ZL3*c!ڐWqZ\cJfw|rh*F!Uba 7۸<|֮<Ӗ %:f%U!}/*9* RnOo VM砈rU*@iI H`L!M2ez{wnME6Ϗ6Mi"cn1zj <=V+n1:eC)뫢}~̗-x YھJɮ~ߒůR#!οu]ǽc3D} дoj}U@ƒ~1<$gfN4`ᆴH*8^7TP z4vߡ憔>TX=j a̴F8LVR) *(34T?kZ޿A*HEtwn$7|E+ͣ%=+̱`aLyVbxPc*Tp^uMд!y|6)\v|5K#ŕ\eVeӨ(x}*4r| &ggAI~+R(CB(&9: sy, hUW1F+5}[S"mr)ǐDD43ߞmQ>-=߀LlPY#y[aV7Z]ܨi.J~C%߱'D>|ꝫmh79uO]wk&]K,.ٴ6oŋwgv[@`eZ;tQ)_r{m*Ӧb[̺삵rkXZ61no%+$\ˁ @C+b ğe1x,MVqwK*J0v}<#<+Nn' Vltǝ7#5uNV_V&gmNsgZ-1? h;*l٤@&(zpZ.E\3Oj+d/p.<3Wqť#zl-P:"&EŢˊ2Dߞ8cnA/_&U^!fѨbh)$k00 & ̟ -<;||7 szmL?Q`柌RF4l>eP̞Iݬ}=Hc:\ݕ^J 1~{SN 7uEEڪDn*}^3m֦2ܱ2R NVσ.ê nK9#zonr 6e&(`}':@DnǨ x =)| !Xpz_X- Sp/&`SΟM\s_X|1 fC-ݗ7tg\KsM})gHb>)*we yψk X|V@S8!-}`bWW1x f@z*!<ŗ4&z \\+w/Nh}JnMШo֘PUK4GHɢ"jp6ͼOC\djjx`,/OKej>))Tۥ57R{o`W{Y3BP^ x3 ^mOG2tewVh  ^[@cɬxpGJ{8.^\ [@Jp19VR^Mduqn N %RrNw~b- _~j|%H鴸ӿg描+'ODRfuV>퉥 ]AgO ' }V.sUudr \uj0$LO?;g }&an+V @;]gI)2k$4JlG!>#X~qc4tZ _~9B4[jKNS5WTڒz hƳCGęy~] r4`6ͪtړKpYCo)5,Vy/˥aJ%5_D [_D8+W̍|Bqeᘊ5Y8NCь:+-A[j9.&5RC"eIƴZ!rRBd1ޔjUL=e1QNuq8ie& fi 0c )֊B֭R]Ycf )GLX1Ʃ~j-'Y Wݹ0,`՜h-:vzvٽj󮑣$0CEQ\9=Je y"r}s;aV+$H ^N3jJvwvt'R+/uRSjFU0IA.US2.{N;9PQ7'{Y#ϿcĹjKa\{{guwUqW(9ƪq4wd-xS KʌQC"{~^$yUS*KXL,L)8- d^x>22uWƖC$[#dHs4Y,HDy %MA\r(JjECAÔHO 3pCRKTfv֞Q1F- $Zʾ"cuJFv6^p[c8||Gbbm:Pl_c,VkՀ3V'{ 1n朲@xxu+߀w`9嬴sqn+ [ p@yNaV1-iϳ6A$ДM vR>*U>sHĕ Һ*TE")D^"˱ b(hEnD("%YzP V TQ0dn{1|}-\X ".I8}uvb:(È u8UG0zb ~]n4f ^`q wZg rTq 35H$]F9D_丩RmdeIgTaMGzE '! !CGlo6Oxm&VeP3îMA0cDm#؂G=$D D&=-xi5޶p;&Dy|u'zLIn>-sܕ FՖђJҶOf).'p!PP@/lN>F P*¹FJn+}кZEnX Ygy>n]K uь]h_yQKĈqN_!Ĝ]x55è}\y |â(Se~ۉU V(LR7Mf>w4z[ҢDqBC } 䦕(F~1[.`:[l1ZDShﬢyΰ{t|R?^T}j,c}U3sZ\}9cfւYyNs9o>XRl| fI4y{Lb^ӷr4џ͞Ъu-ﳁ* im0_k۩-.gZ"̗E1ߏaw8d!7 -[mI~WԳ-%Zr<6,UUXK%Ƀx 㩌rX?ߚN3@9EǮt1ia{?)UǔG1*o Kw>zwwpl?rpGN8 {cf-A'k腌z*>V ׭pK# !Uha фv,1q ~!9C^(Mƹ}5fDIpH,wrLOtkm4ҔMsX.*uŔ&Vrl RRq˜+l)4SQVGnB(!9CZ(,i;03oRMW"5ӷD@2sj7 [2h.nk9|GK#~G; v,~,A^1;X,wlvEv= pnsaLu[2Mpv.c8X ғ86)ꤏ'F2nѰ|:wf"кl^tE2. ѝzj2B0rۋj}K&[ۯςRlHSnԘ8O!]_ 6XTBheңLs`LW(ruhozH%ΡݺKd|cge*GڟQʽmH$X6ris#E1HaZoYR37Q&ԙo^~R=.I[ρiiJSR)8UД6pKOl]G&-.@uLj\)͹ZaW8TAc MV;,N)Ctr jl,"6o](";D*#ΩDM UXĥ\+P=M&wrmfSv,iUSkJ̙%-9Cbс+h kQ"&\BL693>'Z\&' ϋɡ=g{H5ҖycH|:x$ %E@4 ER贔(e6m2x>SɶMR1R޶(ϩah9.1%t EF;#3Ɯf L `IضKZ{/gm+ei;|^W0HvQџ]zU`|ÒA}Y4`&q1ZaZNk yXq GA`j8Bp0%/F%!\Ag%%tV۶X';>v 7o;(/ETH !1e℅IKm@@*Z[ g;5fef%\s?n> BR`zoIFhݻM{ֽ-^.nrk|ZH[?IAZpz B>5iǃ~_"w?}\d}IA_RFЗ4o. <₲`',Q9A X5yE@m;"'))>;)>?>bZ7UU>^SBz?O~2Uн5{aJ ?&SRsE?\S[tss3k^&tFˌ& ;dgEV\BNyE5 j"KJ1d/d, (fFh#1.<Kp%w?4G㶊=Ol,wr]4aPC@Z-h OljO nl ve=C.I mKK;ۏ)9+ g oIc+y~^Q8%x(`S*(,…A4 9/|5y/ސSbUn@ݡ;EIֹzQǢC~3E2ME @$9 %w˵Ƌ\eQ7~,Πcn765kv(|`ڥ0y>}K58j`JG)8C*Ђ-7J7N zw勂:M͕sh:n^!KSE8 XI;RjC6[K젓|,.2*÷b^͸ţ<(nc*h|6Rڡ~? a=}#+-s: C*Ln(LW6)Džı6j_ JEvYxo)Qĕ+W͡\% b-Ȯˡߦ^,hO̴ o{]'Ե+`sCM8 jaTiRrr'ht.L9$AqGHlY4cgUrZY5f0f_ ш0" Jc'#hmUբuxc-02\8J4ԛqN`(zpYXih yqm%Ub$ DDZ'^=Az !~vfٳjc%-SΙa~UݦIbDU(y:P5] o !934s->- ͅM%9FRO߲YAc,?a(^\\Ոq.,6Da2YpF4d`r?F_d;!"=<# ppYKmVJF7!H /kKGuxyqdf]JmļxGS*va25A&#dckmk|S\ CkYKN0ŷ oZڜSSg|i;nRQ?$ p(aV+ X; _ܻ;JG ?[,4 _qS\p!]OUBvk"f0(gI21O=;;Уm4Y vR>xܟ0|W!lGoA8&p=eklua>_ALJ:-jq)+C-6|}UT :jtZ ^;ViR%V{0[ǯ?y%Ro+ 7^S̕f:=ZxѳЀ_d]zzģ.J?$iג&CIY \*]Yo#9+_vgY}àv0<413 dj%G`%iveh˒Iblf\`nv0W->2.}rvooVأu@NP!mMȫb0&fh⹤UX%MG}l>j>x"MF4QHl)]̩2"` +-@y̤(k$K+; M3͓T=6'E6? zٓB\nBHv%! Xro+\v8W8E% Qc,&D9XY@S?.esV KN gT5QN SHESbn8[܀j35Gc @GEaEKDzG. FۊSRP!7Zlet ){uLyBa ʳB,߿.-`" 񆕭oA;Uq{vNܾ0:@ {N7$,N;Ak(yX3(&_K*MҀE m + ->!ɶJ QxTlIx1B>(H.푲4qe^uI^=um,owGV XSXEohCNp#j9~&gbW_Ǔx^bјQVڊ̅BնF!&6+( Cd#s!$-v:UZjeUi 4"lgoAv۪6vϣﴲN#JS A-4xQv1ɅJChȖhMK#,/Aʙ)]ge jB3 Qd˵.dD1}Hv_EbmL5EbĈ@~iؓa7;ްMM} )>!USg >vzY>jJyhs槳;nHi_ IP z髟zu0ץA=[7lJfRI9IEhJ ZVzL'YK{:˱*3 JP"V428ȪPo& NxrQV>^MxkV"=DWݍB7'id|1 csrS93 0R* hsy FKY̜VH1$3k?SK1y/F< L`LKQ~47/)h:]reGS߱e !ˀB#QBs3 v遝ƕLjLTVV @8ÿ́_.h;! #R TEyU'_Z: F__LFcOb~l&6pawfoOsj+eMz 6) Fz'Q:yman5֝{ #f,cZߓ/ny):NT $ sX# v*\c7ΊDѶ<ՓvO PbpއDuzkd1G@@"Ip(n](]GlM#`fj2އP\ҋb{!-v +Ҁ`dtKy~`YiN<Kliñb ntBIKVnD)+(XlZ̾,򽰠`80cD^$8bh_a"I+Λgkr6uf$=^LlUo>4<#!iyNpDaI%Tv5KZ} ~9)yjZsAecU{\+b5c|0uz6[ZgX>%Cr5c|wM"./!V׆`baB3NESĵkO-EԡΦM-K9Wj0+#t6nuƾdhL`UJ-;1 s*QU:jo* $XMDVk(%ϥC k*4-',>! j>w 79S?:$bj\3;l ҙwc>Ta$K%]r[0c/pOp{p~usۥ 1pA1.vJo⚫ܨ$?Ìm(47]#inUeoM5c]ĥoEmbj NzEK]#SP.TLealMpTM՗ Xdm׌X˻J7(\rգ`d ;Jh B&`xd`^t[ދq&k~^I!j:~1*LIR5?P6:xtI1Sմ\p!^KI{VվR* ~?<|'n 9O>O@ Ãȑí!jM51&;+CjB*/ &,/ 7T帖^jBFZb Vr(^ wn-5yh^3D9a]stHkI6s=2XKWQY ̊0MK0lšV[faLA P5A9vAA+ %%2Ԑ>*ӭC= -+%0A) {&s#l vQOP%`(AXE2 .6K?_4~u#-qNPe(K 5U֔͘n1MێsHhܴ:㽶z 10˝B;tk<%gzq(szi7ԗh&Yɡ!; %Cr; *wC17_{,D dD8Ga 1-yU=(qWZz”UT"cqFL]_F7^ϽL8@cT4o`5=gWi_DP %kLnl4ワŘÊ]q=Wӓÿxtzu|t493?(KOǿa\]6-?p=@QZ'g*bϳ7^֓ïa_cUb“|z99?'0$< "w6 (F-9@9$zf_dV *p#GH0͑[)>/s)xeIfUUl t{sN\)Ƌ8HW7LWd.a߭}TV2Tf.&^B%TK[] }4(uۏfأG)ڌ-agYYJ8?ps.8„Syʼhdh^4VFd ?m>d.xgR;-V~vo"⤖ɂJ=ZeAs&%O:>9FuU1q7Eeb{kѮLTSX{,rLvpè6IÌ /[B0JDwJt-L辜yWAuD dЮVv|2cP#|XQL1?6Ɓs%C`ig;4Ndb~geǠ2wR)8ĨS,bj_*M5Zp\;EB;C0X+ϸL;9FjU.J#L>?3lTlh͚^O)G}u g>lٶ}bz~pQ|R'nw.JE]ppzꕋ@bBJɪoC^^&T:/1-QHKQ~g}ԍĸ *+O'OڦS#JSIjL0p32V#/1s!(xBPZ@M7F5r䷃0RED1G:>@utӘ޻>Ǿf'6MAHڎD@Ik9X8UdΑQŨ`%V"ixy殀cOfLE[怢ov֫>!84O0f?i3VPL05e(q4a,ZcV?lS5S ~[ !qEX«$"]pj5gO-ƍ>gRѕ:O/8T=RmKzR9sz7~|r J׿m_ Ѯ ~.*.Ւb0-Px)*ոdwA#NHØtO/?!)0RV\#z:Y"4Rd6)yDsX c `6+ah獎]8յ1`o1),IhGw^yD}_Jfs␄%;48]Z'$A)*R ͙1R6i1zfoOcdaiM bl66X|}퀊 lO 6C_Ǿjz D%+*D5Ƅ#Gb&U!v#BdE0eT;-]!#[: pykUĀ%[b)HZxS'3L ʸn: gJ05(h"5MX81`8E5ײ}-d߁z9S!OӦR::Ǖe Vi陁1˕4r;Z^ƒVu ?5YMtFH_H8[Q7] k"6{kB=jk,bĹh**@Hqs, "+LL gVI91A`31DO|>o)}e`ckGל{I":X:j>B`#2VIP = i7#G dU@ffw ;< B]b'sWj9G:ė::}"dwot]<oqx3ÛKDJwԣiS^!ZCUPR<\S?JDLJv oYˡ^V.M2:@GdzheNZ'vKT9mW0:`Te6]7A<;߽)ŭ;^#v<`m|ؖ28h9-yw D< o8}#%9vj?rњd^+?8'jEshJ K2ZG^bCGtVNLqA3VΎ&6 ؄jto,b9B}/ƾ]IȢ61?t¤1y`̊g~6j<'Sw> tMM/^iGHr͛7Oz̜Ӫ{A&ۜu5Xw-$"ޜNՖ]󔥻hAZbYҲ.C$ktD;f> قQˋ++ u<^Gϕ!-<={D͔LQ\5f4\IqQ_90x PHL$O]>h9TO$qAhrڙ˹w׭ߓ/K'uy3>rIh.GE^њaw0}X7F XۨG{`-g&'& o[%Q#ƤMp+%bT39NEY"ƵsDV5u1aFמ"K7*uOӬԯ4ؼ@ ˸JFE$9hd/Rf DzOv ;R PyIbIn!:.s$ROtO:,-VaHTB[i=JbC;nnz 0*~1^F07κ) =8{/[8@#*VC[AVP{оI`>~= ,z~4kx̎図#Η/~Y|E}k u*+D2D˟LDmdGlS9/ʼ60:,ȕD02Ǯaឆ"ofߺ*=w~ # P|JpRzr q$x-R vE{:rKFHtœɲl~SGh_>Jæ0vb{(zG-{J}̠Tq 4J,#$ tCjfv>g'Ɣc+ ";b'HxV1x1՝xvƦ;uG0nNdwOt½I*'ڪvS4/VFДz]H4DsYk>F^]FRk %E {װS Q|b L(X$vLg3d1SqHd'nORm|R 2Qed6Rmvi|gcNeL;uuIN3_f~G%-#,<@Ӊ[5 7Qnܙ{LV\rU+6_.d,g&6I5RCh] K@cMÛ!|#`حOosѬqW|2.R]TQ;KX"Q"I=5$6}F{PvϷiTÿWίt@WrP QyObQYQBNُ=*eğST񉣝s4[X x]D|X;z͜@*M3\o~F|s}ڎC}n]tN+6꽓69X=۠O;56hB;,G~v\TD8H%0Q 氥6hwоPm"gNZ7b's'=FbsQ`dG9=2D>qp3ɔ ѣn!ch&6i7?lKs54i#.~,ݷ m Gvl#)ͽ~˔%t97DxQB-=:W$K%ԋTxK+1[%Sr]$⓰حC X`9m=ࡐC9MSZxeW|n񊹟vO/W݉X_Wݣ;>Ix_;ܡ5M$du c\FwMY#6Vayj~:<ߟqt1̹fNr7sٝ侨QkP^dc6JUv%RBl=Mя`\`%t)DPV\])|I.mKsʑ͵=uiL?HswD 0s'Rvl( EyAieAۜsHk9Zʍk4f =Թ6Z%Ez%g_i78j6Qoo\^QSs1SnyHa-އF 9ZZU/ɷ]r\ǎH;wD/;߮-r@ZO-RUmS%28NQfCS&JgktII }ivQEToG>ۄDzB|JmC5TڻPttEhdf/o KOgn^ĕ`PVPaU%?kk4CMAk hћ7'Hm*isڠ>1x)tg'/.қIb;c^"Cʉe# vHv$ܨlNe[+33#?y|\!.ښk׿{>9ԫ9 GNl^b! '? .po~o=}nQ3(Q[(u#?{ǜV.)wgUDIwo+"G1R8427ްq#\SՎَwU;q!w#F=Fvt(FiYT}1 Ύ H惨i6-x0t6id_Dʞ䦗9D73ck6EBL_Ev.4/}`RP@&/%.<%1~gHlΈ-V&v?整11Dpf("kjxl,.п[<4")3Ƿu~VjQeTtWe𢹰أ: .K.`miT FcҸ ҫ4HUZ$3!`&M0 YIvR`=4}Shov4tiKH;'T`_>t&ØsHn̫AZlǔJqP$`u/Ʌk9iIJ_$̧,!܏y5Il",am1x:|UIraU ?~UIHQ;o/&vQ?A/R01<:eJp}MX KXO8?1Sڛ7Gtqu~MaKދaNMi뎐 W-1= :nQ=bk2 -#$TAEXom7Ƶ!Q˵7WC3LAsgq6Z̷5l:> h-e\1iQ *} >^yՄђ6FW+YYYʠ1JQG3ܔ{Yqens:"&&S([*y9>1ck>N4q/]!_?yy˄g GʦOُ/}laMφP9o3~9XH1sAƼ$΢gܓ5nʃ$dǘ†Ȉ_2ǭNÞyNlU`p0Bs8(WBS(T\sRj (Մ9xS#٤HYؔO<+>:zBQh ϶ )\&$yL=^^mD(jzVY/2jJ rjjC[@r8IC7xj8xT\{'!:1 N άXl%ͲϾԘJԩ&VmvQ*gc2ZŘi 10?[+捱~XQ;Vd4!A 䒆0fΜӪ`)OJgmNl}[ga.4Q 9p_p4bRpr˟2i9gLWc҂9ipG25P $;r-Kt%v"^\Nog%秖y%n8K KSoP&×ӝ$iRK(s`e<S!D1$Zօ $9%g񇎕Jiup=8ՐjIj=چg+!{jb#v}ls.qN.!j0g0g i ja"tۆB>/ o$kW3d]CghS>/fs?sqݿ?47000/Sneg)p̨r/{:iX?cjߒFax^>͖uչf_[,t34JNmKNlsndF_||r/PNHx(poVZ+_/zSVػCAt``1HsHIʪ>b:T̠0f"(ɔsudrrva^j/`%d ;Er֮SA&g c^M!sS=7IQܤI> i/m٪d6$fY]1oy̫Y]1!G渐SOhVKK'jCJ9"V}NU)SFq n }`95i #SUf!&jMOMij0[~{6!ĭP˒R6'Zr>}@eyw ݧ^?ԡt90Mok&ϢRXT{Me]Z&nUe')~yI#Xnw1ŖSO;d˚A rmfw ׳5qxI9u?o=;ob&v}$ta:NUۇ?r7ivHrsGNSIYOm>4ja$ Aك!$Jn g5 Fi [JV}0 =٦H1\՗d3/RTqѬ3{5x~̫1`NdV"I.]S79 rCxgj,X>M И&k/C^`vE09L )uIs/-kUB8q5@?wLԨ~˧= sJ-#M,6F(_}{4)u|qe=o v m3f8kD{l~Q̃&^8t++oi=0\gʇ_Z[<JbN~+E؟Z96串z&X}vgkH۽ܾ&pz} {a7MʨF@sa*^_qzs < lS&F}_L6z#%#fGALދ| giq4.9}%c<9loWꝍ ,%)X`ũzp[cލ!1:[LoX5 +| |lH_+YdZņ EۡXb]7Ftb}Ug=,_b$R/L4$됬/ C0=f9*rorg. [5v-j䩝NIN1Z[1mWX s`{B?ݠvF̃=qj(E3L `}yaO*훎oMA6`*ȁ`:g? jwzF!="jw.7p^#}bC(yΑnPЧ<6d ً][4paC(ؐnYA&Dڱ>"u*Znq:/>۝d~dߟq~X2ښb;GyG#t:C}OZk2B7g yo8N$y$ FGC>]1rFEAۛ?=#\E]Ձ FpIwWC*&@)zvvITK`b/Ba8V[(J hqU9Y8CT^yg eHK:!R>x#5Tz_Uves?:՛;τj93O\~Rh gESL<$(KZȦ5`L0h}kI&[S,1:Ò1հK 6Dc!G=23xW 6N\dH2,:3mOo{PDJϓ6l=DSzפ0EP| XI@[h&&hP5%FeZtkPOy3$AE~ ߰152V&ԏSxJ&ʾ uI$u~i@sƠbCMIJU΍ޠl BoKMmsɑFaQ@)SboUBW_SUPP3itl:uL >Td+5+juˬ?T^DTZoY˝OѸb6a- zˤ(\reAr>k+(1uhM6hEtuP]LI͂1C :p[hgl1]fbGQaKiE51\dj҅"B*LHMm/.E++JKc㊲x$Rlӿ\iY§o]d3bs1\0j$w)$DdQ#B%ϭ_RBY R֦D XG>VTsfWk$Z'fJx{:qf l&ReKU PgZI)8sFV#ܢ-jcTm6mYM<(^M)j z ]|4]TlBhQu%I*=dj_Y\ pm-hTx,g#{Վ*߁2\  PkGq6r؀^i`?p;npF N/v+]NM!-=LdPuFo ՘Rtj MJ Yjƚ51|6&%7ʰRX$(2rFtZ!1d]::$t,~)ޓj͟;7;IOK~tro#"O4%q#0Sj.f+~߿U^Dx>|~is:֟e y.?oZg:hF ue )' R\-_Ýa&gl(87=ֈ5OW//~cZop״sgA 05q^>7M\FPb_p"ܟkzV}.O (6F(_}{nH>ZځDzn>YMh{rNpIgTiK.0Ey-V-Oޘ Sέ`~*ߎy>%-d{Z警;Kn_X,8Ƶ8n\9s#58}cFP?Aasxﺵ5[Wt"Ϣ>wᝠ]^w٘ged9v~{ {K.]3Pk0L Jz^'%{̹xO 5׮{wcM'Ņ&`'M-pM"׻ج22QiQH MRF;)ݨ tnXCw3C!W4fA8уJ5{vkTH{ݘgCugU/D`c^2{K.\ӂs^*xP^z[ xZ}n(kVًmff hPCI/r+iGKRҊy.b.]p Tp Mvp ]Oh6sV!rE; bv[ f0GEFHؙ`692h$88=5FlwJyXkV?Gl'`f:H u=U` :~.ZjB ~&!%ua<`g N겅پwvWX[mTVd|rz:ݫg춝"o^ws<;hU3?(}8gWXYKq/^LkR BQ3b]ĊrBW̠c/C=GU6z؅ 5gF0P{7 #3yihb.dϨJrŹdf"k$Ѫ#lbSnͪIHXdFzvOktaOqEQee}=Eo<7ۡW'&ąC=vސHtbG'[ݚFFgmh'&c\xIaO bl/fq%(޽҃3F籤t:ۑ\/`)aa7?#o!}Y𒨝'_<@(>q6QP;qf9VCQ;aKr$!Iek B8 =4z)x ;7w5@Bz)rgl&`+ aemy؎7Jmf{ÈDu~ XUؤϱb:]QV7%|cw@1*NɩS1Ri(;uީ}`ccaiYJ?rMC)ј=M&o޴;xwBRIDUfd`DFr޴B1ksĹ^y()Co>δaEE2U,4܏gi|͍ pya2[r)]kc$dBinC&su7i#pi dd"Bn/p H8ƊqnM12b#cÝ9xȭz)!'tb1@8ZPުV}蓜*N}| }U/4*QoW7I1EūWAovw?L3;b{C3I擑+(\-:3 oy[2Y{! [ff6]_,sh=dp8T{469 à(* HGI`.>W 5;ו2NC6G>e?gFO gd ,? foϜ:iǰ٫I$p[?w^,YzU}ګ6 MQ.? '-%?d?`ᙰ}+ ܹ8?] !7T2\6AkÊrYGk@Ն謁:a/k({$m/ "esu_#ƲEqNS'M4~IU lf5 FCfᗑsA؁I8 "8ԑa0VL kޱj[kJvbL.I[HLF0]0]shf;odwqG 6[&NGI$&̣Ԧn 0oZ`ʷ{`^R*5}S(+;ګM衩=\RkfGƼln"O>Cp<{qgfԝs/ϓ>^L+q~{:]$^kMoBB!p@lJ _:tލ}Y!Ւm?XCގKQSG'O0WGVF3%JC%4d$ 8T\rzT5Xm:F(z oqAJ7YUBjEFA*zv:D~rnڲhLN1 qSUkӂ/PE緯^؍kfk6%l^+ ؎p H(S1j_0ĺ)F́dN!4=d ٭ezHAִaTB#2 ss_Z2v՞N1X cK@lhe~q~6T G\NIL/4P>q9,. ?Dgk[ٝO8T>p朓HUn6C,Om45P(U24F4<:ҹ[=yIJ7@&O$˞E<=[VzjLGO QgG>Q/H: )R`OaU^IŘԊ;hU0h>@3Әh|X8M )W(d lJB2F,HD2)Z ްQ{TRLǓJBռ! c" ; >K+Z=51f~>D泝c*AcrJ8^4L`W0 i9RV=bH}ȶWcy5Q"z8Us->^V'{ءNH_ =|>#Xiuo9|ï]Hcvۛh+~Q"-Lb TUBb,(>I;Xi $*dsw*4$o__ϳ^Y^CW0_-# ~zH8Ј 64ilBa17Xǩaq]:/=i2sW#fL*ldJmKu-C f8Ie1̐'( tQ6ZJǰЅBJogV0'bλ ?a~λ"OZ=v |LO'1QLVawlg`-ȷ#[T;'RyG1{K|ӈ^m8k\`gjz+!.Uc !&Vq,8R%͎O/ooiڻ#!]aZM0Ibx&8 uPZ0∔&~T +x=5ûdnbcb!IxLRL) '#)IbmL':NmL @RqqV䐂\󄓰KʤPTOS8eͣb\0NRp#R@,B6sw8IN%"x{^S6߯WD$&i`лuʟJUI“PU8hLB0ӜGzbܫBH :V2ؑD c1&% S"76 AR &ּ*$Ev%2MJX1ؑeJɯ>5WX0"?q>:@:96B|CxX vฝ8l2: ٳ O f: ]Su w²@_pEooJ y-x ey ζ w%v#(Xh]|t~ \`_3+6/zhhYm*g zhr7tN_2v묇:aAbJ,,&<f1w? 6CϺgз*pGؕ~$(H]]Y&l:@S}"w{^*z25aNqB1_ȻI/|{?Kk~/󯕐|bFpg| $%TG**ᤑvh*#޿51~9C̒yf88s~t;p` S5iD#ydqS޹IR' uX VZan8"$Vtp=([e|FJ4yiж[zq/H[#nAPG+^$éG22vҖ<~;L CZ r28١31 x03w?otA2/x}%*J}4?2"8!A"8!pDQ̸j'?!i6gQϥ .<Ü V>\q k&]J-fn]V z}g.fMK\,w&},TI,K\#K\ő?M4J2; uYmO5B^E%z(X<-oVE8%tݹp?>CL/e:AџZFS`%u~]vri\g+יؙvx: +LJQ$RIC,ʂm&cFJ9Đ]F3Y,Y], &{1vc3bYbFKKJۚm<M0][mj&?ndRy8`"RkXzUSJ#jbRAY[*HX*T4I,pG*4/la3uyV,x8KVkJi8 zuB =nez5,Ugv`+NHLjY뗱Uҽm'TnQ?!I76\|$si*R2NѕT|ƪh(M$KQ" Hxƒ) D!%&3 >z?AIfOs#:i]_^?ֺĤi$upJ&` 0a('1Ů.$Ɉ,;.b\cJdX3mj0  &&RfyǕKW󩫛YĄ|e6s!wq'|'ʻo:$[,_WA< Ր¥UyP*e) zpٟBgٟwFf6aa 4eK{+iĀ@ IQ6@-4} G`xat` %j"X܄~0> ;8h#YP Ch,~(gDy'$x4; BtX@7/^Ɨ{ҝ3 C~]7Vgnb`=f`qF8dԄF+R Ym̪Ə։B2ɣTܷ#cK6"M\ң&,Ds-snDw|Qӽ-6SUC_%M';7`_?dU]6hݛlOLvmx>8G y Z%&)&)iD BJb`(9{uWwפp {o­Yxx>- ,Rj~u`nʛL>88`?,[XuBN7?p4 t5{ Uz7zֱWjF^ :oK0ٮN%Cۃѿ'uuw,rLҝnOo _<ґ5>k$y?{Xe ʺ\h4/_up;} R=h4MwTX4ղ_ЗAB#·ӱO}kC؃i+Y7}A~&{]Vm߶.ObZu46jE=iGZa>  m'jz%ɶ$kRjJӼ^W"չ Z+qn/!|~ {]R.v*D;4Hd4%`[M kL׶2:v7̡kwjwlg'FJi3m:6!9}wEOC5>"yaB9GjyCݩm8/KQ?8Bn}v@nA;-hH#N@-G3Gs> o*%+ާׄ,?f8˃ykBM!?T6&ПgW^z*(N/>^_aU(_~:LF:#)f+CJQ "Cd1!BgXp=R.5mH^-qُͰ0ܔB_,]\WNCh%Han$'&&)Fn:eKռ0gxΫwi S/7Oj1hρYJ5bS(OA p׿<{1jj=q4^ 95>x#g@?}(8&1 N{[p0EZkL?ݗ3/F7Oz00V?GÔxSfіT@;ea] xe\#r%GHOhj~< D:}w[0gpIx8fRpY1Gh,B@rV%%34gQ`9@ɔbn5ϾhP۰cV@ׄ ayTm&)9YqI&Z 49PNh2&keFX$WL? *JbbT[ݾΚ mjC~1f:wJ% r܎JVB61(UB!D>zj0=\+GX4 t{Dh. =RqK 0F Ec;H#6T$բ6ݚHYx4| Gw!F3#IIg:2-p x'267yHOj-,IuKPAi-YV$"A8 X A$^ !(j'#0'} g\Kz=4zUd/tЀBdt CtJ10s!8k($VyiHВQwUd]q.\ !"-h␟-`m-%BL[De14e |ŒMZV-s!icRHm+;mB~c @I$X$Q.w '] !RV˨s8jYt_$T'0CГa4s&t ѕ-:Bf+1=I>BjDLJyO BӜ È}QmHF1֒P 5Yhb0lRуaAXJ1h+2,T9CZbcY{ x$\.*p6ҕӠ66 D"P"Mf d)y0DjC P\%,M Õ Ic#&gPPpRcᜓʆp5 r56C6-5[?*ex V_gIC,.-JO|x=;&|7ev`wOTXTx|U‰]}g(W rh1{{0ΖכˋoNSFFTs馳,wK_s-\77dmQ] xBgNKwޑ;-nz=/1f,n4ӻm?9MBb,×GWmwjS(cu/òOtqgcZg.|Tބ=żݛCi{J>~' 4>HP/SXc\">[Wü(OWHRFkGzEHCHLmZѡ€R+:RC6Ji&afo:ak<Ylk(nNCkv&߱C^sdֆ)Z3:֮Ŕ {tw3q{638 m&sG"k4x* Vv@CyMG+MY'3ЇU9{[X39] mKfTWS==(#܇x9rU5'^QZ6OwG>.Y%~~غm;QI Q?—k)+ڈfuꞈk^)Rx EVL6N'bfH# WWkޕ[uFY{mҜc9$E e'%Ejjo3TVg|~7<;YgS!V#)1Iʴ~wyu^_3f̱{laun!~k SEx*%};RQ؆1OpoVf;V,ο\W+(0 P!̆W";=3jes0eڿҨEՐASXod*>ÀyPQe^զJw۴j@4|_v_ S3Sңu+yY[ѝ۳GPy89nv`ldF/qgpBفsErܳIݗ:nPw!`Ca7w{ `: [ʚAyY̬4f㓎x@qM1]C7 ̀юuUZ]CcnH0b6-{V߯2 ݋ܧ6G^ǓhΠc~|V+ʙrsS"<ݱK zf.7o9痐p},ʘ*6rl^8ҊI͕.H=!1ews9# QVeݷ_>9F)~.:Uy9=}X柏y[$ofoN~]>Q>BRy1;_>(=YP3כ!H:p:<])u"~ŋ_?Orv(i9VrP9FABejitdɾǽS&,>oM^pb 0"c98N 6S,ۛ,'>Mq7'K9i*=;N6DߏТ@#K>.V 5Ko~e`@()g7D1W\_14Ӱ <4k+Uo=XTG0Gp}7kή`)pG3 K-To՚GiN}@G݀+v1%gTf9)JSh@UA_}tx5F<#QpF,YP)[U?[HWޕ=S:2hY֫Wo&(6~@Sȳr yVN!)Yr0#uR8$HL Jʃ69pيS.f܆ѿj>yrJ}:z7l\߃h~$0$nQby :нYT#Z AZ[oimD@~6l weBW,DOhH4bYÉAQAbr WYz)y0|'.4 uR;H7\ w6b_Z&EJOPÁ3ϸVG=F`H,)ɔCnQhrt^ hEEvG#tSTmfj*\&EP<*`$Es4*M,@JiR^N:ESr@rʨ59db0>Dn¿^:oxu5 fxȈQX _ ϸQSӯpr{8FbT잕ݳR{Vح9TjGr)Y˔W^19> #|UNmȵ|gQT<-tFqz>RpkWLTSRhyeTn\fi侢ϣBRظv)6튍ubFr *^Xl܀vB a^+`M G,RnH ET9ѧL# %EґFaZb5 HR'(j@W:L?{F俊?nnn6r .C$fh#KIcrW,GfKv2g&Us|هҍY 荵2R+F6fCTsW\qsɘʯm@#ֱ)6S۵ ,%!(2$c@/S.|԰%?tʙK$*D=cND }UeRk5g=9[ԭr@< c;՜U+=KKܣ= M9X2eu0L\ 83%0;ƵTa.  %!2gZ 0],$ae -%V|g`Z`Gc:VKµ@2[TQ-YiA4֌о d8HF-*!Fn68#$=I%08 xNocv]0h&1nqu9Ѳ3 .Ld^Xsb_?__wAj0). Q2FF i,P՝::v6 1bw!*``=~>Tw(CYh)`9E,JZ1l)orkg< cZ̯Fk^ JD%C; FFw!JE偦ľe2LrKJ6u ~ȧ]) i+>!` ;@FSI̡ky)o/Q[Zg}AXQ + w"wiܥYrPYvT*T&{dS/cn&,"PEpQY*ֈh%JHl0 /7vДڻ\bEO۸t6.KGoKGڻBzCR%NiH\3 ω&Ȑڻ/ʪ `3PD}8]]t-ҵ6p>qaW5VJ %Ec\.+yk P ])aX3<T9 SYXKA 1$%2Hg)*bD= Jh7j%RQ30?XSTb&debyeFb8)D(h@X Y+5 jVXoQ4ŒU[JI'pqg>uhS):v`_((˭lPl3, vSPbW"p-Kt7պ L97Gg&^6T"W p5ߕDUKּkCJS+X[MEa@M Xny}h+ΩbΖx/^`l>kHna<7+3e!X6*d_#φ _ i2g )tFw.k)=Jz\((2 )+2-M7ҥIc_ur9}UZ=0p'19QJ7kqjC<}twlKBm5^"aŻ9Yh"^ }>;Vr7fp>5 Xb-4WҘ n?< B91a#v:,&G ˒:Xmԁ8FtZVD 8brϤG_TcDQgRhXK= 5g0Z왯z*Y]X)43;-A|a^h,p7!ECE2`*8+gZSp׹”h (g89]Đ9' o[%Byp}FrdC;7#XK_}@ z䱮zt},̝drĥHN$qre/^YLp6~:]X-aFiUaѦhc^[s9dH-ɀ~؃ISԨ++ݽDP"R+%t?u/1n%_|޺N -I;'zvID ̎pgmhYӄN? E kAHϺ~:BI4T?*y"H 5Ξ<n~ odPaBZ)։Dx|bc4ͥ A*k@RK5ueI,e0qhL:{lOPb}?i<`h=/+U N VBճ="TcqSɲ 6{ n":\\>:'zT"AgG`tNgQ^\U [\T :k32_j=n/Å}[sw.|1!RKΏcӟDr3[f OlDNQa49Dt~b:*3XB]P?('<<6c’3 STwO_>z,HazFY%guo!wyLb]v: Tsk@qcG]w/\{U=#{\QO.쟧p|Zo&;v6\;Ԍ@'C6a*DAr8'Wތ~V0G3= b8)ǣOg?zFg $WEfj£ǯKs^I޳pK@W_96Y*/=[Ψ>ӋzbeO j6d4~}0fǯbOkPwx= f4AT,>~߀yS=Vg2lVο[uM3ܼ81Da6c5<\@] ZeXj>)`1]췷aja wP:` #`0MҚX%oXA|68o1+#%k-X3Ĺ\dFSmtVqbZ:EMpЎwKwaM}dfE7GhVuRY9rtW],)Ԃ?oЃAt\>4!) M''O -Kkgo6Eey]˫ṕEҚ7+-eC4vqD;L8q6$x_^]df8Zw7-Di{ Rm%_8QX KZw ϛv#az@p4d!\i#G”TR@u^8dۑ4wk: {1ƺ+2Ԥ l mISBAVAjPe E8U}PIK{ws zk Jۋ@ZM X>t">f5?r@?{6쿊?ܳ'r~(voݢHc-)!RBْ$B֖)p834FyJ5Bc`1:Iw%*/eZ3n? -icOU]c&츯nz]UCQΪϣ(5Q ds]u8|w4WvdON:7kZ2- ZirS޻ #is-k!c8uڥ8>8ݸ %UG6vE)kol^8V$ZzU -~d.gdj6&HlmDБE@k'Sb>/e^TXЎpmJ! R%xH\x#Y%ݦKc%n/m/?\q#.rƏV"Z sT@h2e>]{?% #7(Ir%v%ɝ%${ &@£MƆ/gxHA~h?Ĥv cH"XXQK-J!#`HlҽC$ɶJD@UY8HzS&A6|WJJ/ h$qvm>W:2}-U / kAo^#ܦ@Qh6mkvn#YC6f`AXP.V^Zv`k#ѲT}kgg[X[aiAH=wZa3 b [j[@seAU cD駟z{sALԹ(dD]DיP]{{z3g<3W?P =; f 0y?uU!oUM̱ɼ3~ӯ|vfgOmtS 8{sU,PzJv-RM\V7Zfg٥P @\Z=6V"$X]d7AsW+phil(bJeNJhlhᓫܮ ,S49McW"Mw 4\~&wkH 0)ՔZ@PBLtmlkn@jhSo= =zX '%`8a1Yp؉=HG #``bog(ǔK K,!,ۇQӳO$WDw3m1=D]c$m>c]'s 놓4>sBZ%I#_$`N1(b+GwzSvB:#A F Q,9W,ycV vL Tt-a<@Pk@P՞ ZN6ewc383K(w,(3Ke/M,,AH._xm۴`Q1fl>wRE[ȄQ! >Wrz呄P[f්r@ ƏX:0!1W`YG+UcHďjO1Ov.nG~XmV)$x>x=;>-K bWQq.~-H!0=mVBø|p8*OB`L>))db]Mfg:pZ ESE* ˝~QA!$Q06~8+_6~(ؐ%L$2#h>G"!tu}Bsy T9+6Y6a.,C0FoЧ,: P .:)!_|*TB}dKTv"̥Q DZ4Xl R[, eHIΖ^bd;H!8zG3n2?J" (ikY :8?k-i-,@?ɧDtGBsHRn#K^6H5XA[wwVolB]8ͥ[|g*#֞k5VEp$[M$+ ZJ &")S"HP 1a¸U%CCzRY!R_7AA|7 J` ;,9h석!Cu a!vN(vL[C.R|e)>|YI! 54ј^rCm ԚHa$DVhL #@$VL NiO?L^v',:a}Q6 @%=Y铫3o~bq&({?U䴍65Zt&n `j ss=+F !qDđH<ƒ؇e-!.vrT aM%$nXW.q4ΎW(D8*XSۏk3a:@bk)gU@k'7; 9q03 b)S*N 5 (ʗf/;s1>J"q_;SU'㲳'01t`R3%&dPЖ Jy'w!]v7@%eXP{GCiڈyk#F6 hh@HEu*:`*ĸD"!"&Śj)(14aPBbJF6hW5=w"/<{bK؏M~vfh`Rnnڎ{MFY:jBf368k6NOflӱ?`€= Aτe Z]oqm/$OOJY6f0h)k 1Bs3F-bT'*aV=}FR,Ԁ:M%Mo7XScF@@ 'JV)N<%n]gq,!̪kqy}T5ʕr'Mq,^sBhHXlڭyDHP AP[n$IsS̴aI L@,N0ᄉu(Oe.dl?&@Ul$`hΡO|.q|hF|hxsc H c1< %a)qN_K #Ѭ-Iu}!mH)mSbLBZke9 ԘXaB< XC 0d9gnSfp\P@-0(B8-嘔ugTE`aJ HwқRvJ.X3-d/* Fk0Su9Ez:qtsґ3B d\tBD*]3AuH$AA5b7My" UZ:L1J '`H˕EbʤuOtZ(@m;991v`4E(0qm(hh,\oEg5(YƇr$_>x /|>NU$& ˬCԷ/_-1aZIlyIEJyo;gF^]#%! Gt~ܱN23<}⣀o^7@MgY=9\`ATY \ݹx/y;\>B_=SL' lb[Jh3g_gV"86D =Y90\!ml"[fKOBhIN7"IDv*g7):+>VY-r1D Mgt4Qj)2W cJEbx "mXpЪʒ$"DI uB]B0^!m|6-䣭{ M$Y2rcH\bE-!|[֕zwXV>$gQ(2-([L ܙCHJ7B)SvOaLNYCV]{ƑB_rءv.Xf7H.ɗBOwX"e/>/X&- 9]WWWWS%‘)4x JPV.'#G lmC!LAjG[{x.Qn`4,3v(9lKS 14.+,t$;y^]:- y.'̆ԿFO;ae\x;jU5rjLAJn\`dQ  ^-0̋1ϋ_fk-midmN'N#Vj8 HYƌ MgLpH\U &%"h71~Z(l;eqaQ- Kl("%!<#)a?'y`f.<-}x8@ |=*\Lƾ1aJ:#}L'Vi)R">qKv:[XXTX j>4(]܇3dn\"ZBoW 4YH *~Q!Xd l6khr{ElwME Rwt 72}& cZo^g!Zބc2Εk;q~cH<bEU P+{7J?~|y!jR(7b5)}7%*:oWXikk+puEufk7U- n.MKKTtkLݮ-? \Qb$(&I謔(̼2O gQun_i90EAw_uu8 8_2b- ʬ" 3_oo'ǃ#Hw4gzi-Pw "Xq3}<^ž ?=_`_}7rGxJwtQw8x*KeZdp/4qpt`@4S +ͅ+XKv\IAuncs8 K z{V>ZB36Z Pm9 L :`M$Qܿ\qQD ˯Oh;yV SQ%Qob I|=^'1Kuv~E;rTN}Dy7 GjaK. pi8rSbGP;\;}*qKbrx{,A̚ErL] Br&&lOPҖCU6Q]:~r棵@O~WAw/n$䟵>U$Ze`B-2A*y ɇ?3-޸Gׄb0H>,r4,@Xg*1!TtWqat%Ft(_*1dixIY7`)ZdD[KqZW꘻q@>RoJvè=t%ל]Ryݍ~`ث!%u Qthf,8ַ^!_3*,q>Ĝ%5.˅M:-imIBH#/wpk|`q_X Kja6;1tXW4hqoBsk4e17Gf$}x?tS7M7uMM-NG$M-*+ 0;$usKSN%rgʌ3TSn)ai5_5ŬkŬлm}ৼ!6`ꊱ?A29Ko#W;(fhgk/Ko\L4dr66;CWqj4&^YAΰ`Ⱥ,5B+vj }T%ݚYXVaw媙֛uOf{xjLPwyGvV_Q,Y6m\S* {,0|:ōF1F N(I=&X׺_I霧+FYCRYd8#"7ft7L>^^@j'B eІa=?I6p}WUz-=28I3nP%eR.`/ظ L<ғۦv;V Cl4K$4@Ah b0H5Yao;qCN'j2^\3|ِp DVPݪt 4ryOPIXR"AeX4)QhN'Dkpgaǧ?31PX|G,OXRwWp:k5q5Qh&AI[U`hzLun'yҲbQ%߼TO0^<]٨ ,^5Z>Aq6(lYM[E#XjqK +-2iaqa5Σt_HDM w@H!˟ [iG9kbptiKufe֤k O֖휗Tb,y`HSkt:q+,_HR̋LSL2e zT"sRuDX* nCBjSW8){ @ t l!7*o o EU6sj +,gP(W2eh*BZцk1P @Z74 .@OiY߉ 7۰DFgV^ +\k-@ YJ8󂌉|j2Ua|`Ǚ-)y=8J48zCʵ'|A7.m[j#\Jju&!8-_h PQٝq Y,1~oߛA6&iȬ$d]ZjOYOqUI׻XA!MB%ԢKsٹZZB]/g^L>N72Ɠ|dg?`~xI!I ܏vtQc/U 5pq1n9 }*^QmM/,DTi&e"b78mKYv52vc)[>7kvlrOA )j4m3+3#Pa:b%=N'SOfv-܆!ফ|)r/X)"Y.B\x JK1UJ3 3Ӟr/RC:Ty&4.F6ݷeC 19Pc a3|aГ܊)r9 v;hd;C$j`P}Xvwʶ<C% .zW>q&KBl–Ԁ4N Xb4{S%ʹ6N N9$Q! ^ E*]|96#8bqݭQB2Mӌ')N$T0hgY ]XHәb{ 0`' {"nK"AIvo9%J,a"М!Kν*<%)j-KOw2# T|N1DZl@*k{d[8 9,M0"41 lGi"IkMcL*3)񘦙HSf`恙\!vgm?lr.u Lc*$NRxGB udjcREEihCHfS VC$Y +RD&ل-2]\J6.{1}TsnxZHǚ;3w\u:ZuyPǝ<7qņS+AQk? \w?A[hr#Ѫģw @+អoh<BU{RuO `:4}g)t9;YӝC M(==7hts2΁(gT-!zj uof a`4Nۚ,AgCDkm#Go{ؖ(qM0O9dtk[Y<_%-%L$zW_z mVGع]Ogzݼ4t]|Ə>f2l&jqL>P228*F ݨ3MGe ;gAGe){e_?t /Р,jp%j K}>l/HΪكL4< . cmFPC}@taP*-+f#D(YQw^Yc18רaI6A{F?,1!6<Ѐ}xh/(0k(q =xrXQ;nsNYVAp0ec'uc̆ 1.Vк1ԇ@bJAíe1Qr'B7>hE(2\L(La!MHJs{_5~6艀Ut'"HKNO~'N և1Z13IAF C5Ŷg͸<15zbIELMGS6XJFsS,Ҧf)ra;7&A&c)"M$Я)xL6C?|Vi+N!0?ՃʄVŒBg [LZ)/Y [߼SYB&SKz}#ʾw`a[XhX\Q` ; a ";m}PdB}4$CQ}>j-RUe}1e5MemCJϻbB?"CRqnDgQtp94īHƂFcVByA~+ڿ{(y'nji5oLp/3d1xF|lXwoY4aXk'X_ޫ|zU Q|[=Prz~[7ɚJ#U.:p8ϒN߾{{{q~}ǧAHi(,p.SBvpup yME#Hz2@>8sDVLށ64Ƹ,?~ceFi$#K'ST]X,ost ͗뷼o_m `o&q>\Vt*kյ!W=zp}.]brȢRzh=|n0AZE%/||ϭ-,m񼸺ӒH,rP=j kp<X6g2Mz/ v@-DӣD˙yO-g2C|h!c?3{{D@RjuM2B;%7CG] Ksz-cG6&8I? xۋo+kH8 5f'kiPRE/5CVp^"Fb Z8NDOGNh/"(#bNQ (#Q$H@K:*w 3ĸ]: 2`jOڡQ mq4`f90ЈM:Ĉ,Mgу> TЋ+X&ÿ]S?rfm b Y^-cJ_ CB=lDDK;}Fb\@8 X4dd)_.Tw B!:|d29y#3x/.&%3^p)5y .-eّ6af"ۦ06&t#2#XLãJC@K P hD:J i&+]=M6ud]˺@_sfKuDpL)CڠI(QޢXy0Y|v1$8LSWX4ւԦ@)K%05VQbEru9FN3 h[%DǶb@U4NUdrV <:eb !eذԔs$Tl(Cf/MLa<$$9f=-щ,S\j{1/!Y+bc 9 jhRA:6Z!+Ar s?&֒ZLOF tO{4!JO~d9,Tk& zq$0V1M:)zIHl`;U1%EkBțWgmtW*# hHIЇ2[LPAF<$MzGvk-!kR-D/jkP~T7\L,e]=40rۯq)_PA[=-u(> mx&'pM*S`,[uOT`ILf{4lI45sCz\< `>PAO@0 "mOPit̨1u]Z-b^bkrA im` CVll/fm;N5 o`\eh +*"}Ҥ^Y% [ÃIŨDjhzP}_qYXLk\Z AUp%lTp ݽRS-ضe[b`&199.\`ݮ 5W?陊KӦ񚎊gTf㭚}ۥY/xQ]h^AuIc(t"b6Hw_hl<ՎAюGџG \Sho(u EYn}2ۛZpW۫mӁ. )v$}o);01`4|}]Thy]dgԺ#}/,SGcyrv^ebvA}nzSf3#|46'X\F s0k9;Q 4dw v~6z !'~2k|燛P. g;F%_}/߮7]tO:Y>./ ~3CZ@s$i [%KyI+JT)*\ 8ERNv`+U@>,G>!ͮ'3_%"qUˆ'`vLfQ3Qrx7s .Ȗ/Z˿?;U_\ǹ=%Eg$bzUjmVpqi[4@c>Y+.3W*\=l%q}UDBHE~O% $4}l?_eD5!G=D'A2-I=C HDJ2LC |M`,|jNCEu9;ԂXrrVזN_|LSb%}"图~~Ǔ77vQ}Ao k 5t-~e|ˊ[]F߲2OC+Aϻ酹IKy@9JA6ׁ4Wb>dۻ[.N;x#fV[6]hN>ժ`˻[.N;x#Vz6alMhW6|*S~3T]d#0g~ ~I;.N OK徭"Leɟ~7?'18IWN{ޠߝ~ wziA|wMeMexE;2BrϹ"#!! BjcOGWHt4FG@594C-F1֡5Sd(p3 &IrT e :R&5II j^ 1!GU)XmJ'句% .!?0!GpcbztlRVpƘzMT14IՐ C.B3-15UP=*SaHT3&S#z,I,Rlj) 1SrqK6mmHYNjǞѣv$È"OOUbGq#e՘xx* rA`JWe$3s锢K@@/e=Z#zuͥ_U/U:q]"!O_ L}h!sogM"c3)pct؜nObAF̷36KH3"HNnJ9h\N]toZ<ݲ Mn%$䙋hLQ<{yGQnb|%Dt/#r `o&\=}@.)㮁\mW((f7"uPstMgN]SȖ)^ +v؊Qi>lGg|r[&Z6u9 vZAuoͣ%co ^J}vM 0Ty5]9JD&A8L4h'T!>L)@4d$pJlVN?S057 ${2@Za~tk6-c̘s>^ht7YMOMHۥo]KM耖dQmpGȧ&(w@TQA5Ϸc~=:ev"3b<*~Bȣnt<JT1AtI&}}8Lq ByX &re4uK 26|,5!~uPRE'!o% EZ@BLZlX,@0lΘ5E F?ݶ]  vLV,pSiNTEֵD{*+e8h(A!CpR3nG4P(1BE=lM|reTVabXDj!fI.R_"u(pg7|\k2ءL POH!n\@RMyМ9c1!0֟QP|*ೈAJ`S]bjLA{or&ޕg.92%Ʒx3z>~~3SBx}r MW@Bf[M4ڻLry":ctnSz\xYaلjEL\`@5dHIھMw])guUw3)hΤ=8YSX͑͐2&15ql|9gzWUmGs d_vrQ?B_85F'HfR(n.132ERmt0&D^263 o7:(ޮ ~ !6a޾`0U&c"b- Ui<Ύd80l?%} J v>P.į"6GR(.VK eT"J RLXv{[i3X;m5>͘3󺠷sE8͊H-3hҺֵjU[~;[^-NkSbyZW=skt8wW-7+nxro2\LnO_[>;^ZROB2789T;MpZkZ֚ഩ n_~U퓕R[&?z/kc?rsYO6VV׺,{ݘ/& 7TK_7 <疾j^]X̤ʾ- 7.ozk./eiox*YRh[P34(Đ0L,\FFa$!UmDQ; 6nsSG{v" -~/X4Ir'%=1pj?ݜގ o>>sAZ򐊂)Ci[q0:DRyTмUg :D> }H/g./VeǛ7`s1xusW)^aF B  BrY(4;^Bad@0nr(jo҇]XxyX:s5D쁅aN Sҟ= H,H˄dd@"磼Jl3eUbN͗\%ьwM**) ].SѲ$N']J21RTI( IR%Y\?r|*vfY$S2D4֕P ̊92JrdMn*Hy9栓rMn/C@65S[`.v;PNL3%}Zb(H;AXB D5T~z7V@gݤ27g8 {?oחiï9eҋH+AWL4âB(FJi5SFYE_O.r CppKOCU,J}!J= Ė1>{$?1 &^Pn+EbdPV{t 0Cb"#P؃6e+@̏ DN>OxٜB) ˦TfCbqDRW6SQm((O#:K[wR_Qi.RO; aErcJjJMH0X86ʋT9a{b9 3N}Ys].DY-l~!ۃuEIQ90E ӯ)lH4"y0_[FSɎ[]Z۾FG+ 0՝vZlvZlMkvu%3Znbg9E), $x(0 +L^?^;z\mm`V'mO~ޙxٍTix\uq FalfHa@8ӣ"o{es}wẃ_8c}ejJ U|U߯(1qnGL:Fuq$g]nkNӛcH6Z=1?oVݜGqWG^m~\/rvy[xsy'on5<G'GoL띝Rȳ߹f7:Wq{\T.c;`]խU ?vLq%P4tɫGEtU`^`d2?Zl3~m(§A^QVv<>5?Üٿ.>0CJs<9֍Sxme[vqV]~FMgiRy+ BDRƃè ҊV©8^nEsēve?zk/ݧe U(0a#C(h4(43"(EH9"15$9FOgYjQPiGuȠ~(pt)EpEDtQT#r֙NjBK */pĘdY:\9"dVAp^llexBڈy L=ΝEzL^`$aSF3rtgB)4b@Gd8f9YaB)1yT ZZ˭<MXBSy6AuGIp.,jFJ -Xytp(+')·.hQ1Bax^߈ۃUAş{Q%E8I(#(Dm̸i7=%^\~1ILIz8 F ߒb$6rm.}jpG{,Է4f ,'f>mQH|${L0vٴɇj$nVJ-;~IHo[}ٺrM>Zk(!X; (\RFRq^\,bw*Ե0U b-b&GeS,wjr*8 WHi#0+/PmnC<|Wvu2ox\腌@eFWK2.gY-B.–ҢUKy6K<%dz'Zz'ŷ,jɵ ޶ VYhH. =`t;چğkH4<#N.YdHYLBg)CPNO+|' 5,Qe;8M8/I]'2ٵ(|L2*k5Gi ϣQۍ64 cj-u8ORþ;Y:;|ZA6`M}{-خF`I]ݰy k|{mGuzqˑv| ?mw6~/`ο8f+&v`TuًoƼk-ĺͫff;c[_q] Ɨڶ_+i 6jN}x;]6Cxгq5JIȠ} A{ct^NGB# Agn7!\tBE),$bR"HaU)KkrnE, L'-bֵ>Py~=]wG,-;`XI֢}Z/H'h뿙4sE ׽\N=v뇪\om;l*k\}9>@ C&)wfݜ(hD觹'o\.Lcfzdx'aoЏRkOln[70F.O@طR|Që9xV'* = kSD#aI$qYxH*q `1Xk]6YL vb¯Mm#5 ͧv7\ׂ?ࡱX,o)X@в]=j?Os?jeY8[i?ѓ~88Kp)j5I :Zpxk_Ϩ >AC͔ϯWo^|yvt=tZXj.޼zwvu͇7/]]8-˨I0K-1jw޾{bi^Xof*PO+`k]ʯڡ9h6aargϺ/J[9vӏ s80=ޠ+nwY6}vƫVRʝ)_S|On9ܥn>ִ)Z2޾Q+5^еf|Qm@3 &mb x(<ʺϵ:O^h:( 61pW}*܅!,/0n_q@ҡ {C^F[95{o~h^'͚=5J}W?6OƘaNfG*TA 2ULd ӑCBZi{I#ݛ'_深W\U0a `X*V xn3sT=ȥlJG\>փv>dqyq}iF!YvH/Sq_A{Tt|c}n껯)R]k,wR>n\} vu|lvj YĄcdFkK$Jb!0HL1vI!/_~K*D>c | $_&MOk#(Ђ1$1 FH'2!q7G]#`}>Ff,zp@s( n&ZjnL|i2N ZD\ Y"\aLqhǡ9DaEe4%c %I55)xEBbhXoxPieUEUȴ"d"g*J a!vxLd S*TA 2Uh QX |g*+V U0a `X*\LeuӇu,w|p}g*b3TVg*3=S) &B#cUAUI9 d* LyI!d2[]k8ZQs^UCmUZUd?c6>)r1Zs֊<(f+am6_tYPt#u+o-L;f_)Wrϛ,g ..ezy z6K;tnR|L >#}>uA+ܾܓےKK 857ї*ll󃨭d2[ ?,ؽ2>!Ik$uÈ*wrޱ7!M<:" K4aox(E#5w?7VGeX{pJ3dU;%_ׇ=m "#mS6|b z+va/ ;cX p`ԇHõKI%#FPHeA"26 QAd$rid ETb;c졿R2tfJJB1X!ELő& a4a8dFX#KAX$D M|bϥiv弊#XܝwLnoj$~Nuz8,ue>̰N<VmotG(lQNcm6ҾsQ+Ćzo  "^X [ň cC*V*JBD&5Gu _ΎӜ䳰ѺϕyjozWdQ WHcii{>Ej1m-xܳC Ppt4T~T1]XBũ t;*ѠF04wr>QNYv4䉫hb=֘n6A):pt֭4E64䉫hXdκ 2>~VETlXí3L%ٛu+MhQ y*ZFh.zX.cneyPDuu;~)S{n -jݶ!O\EKG9oݔ:iYQ}co 2չn -jݶ!O\EK9w!sMbtح,N>ֱn|%EvmJZԺmCC.SΆ\~i:7cӞ~?.#1<ct$6['/)Cd )%.]'7q;dӮ?蹐e't(<}׵n '+k=s1\-\ĎϾ!o./4%S.̐.pdcTat(`e"#ՈxkO;; | )8؜6*\Zq|Ƈi%ˇᥚtg|e\F*) VUFdAW%.'bs ݡPf:ۢpA `HX5K6K֑οfɖ$@A=Ttqj~YbE1)I nwYO;}`M%%Y.VŨR2*F-g2J"}؝zF Q1*YȨ&=gCKuXdw:}f-F?qet-փoNm C@9g7s-L 2c&hT13eZLVI$gTTLb /9Ls%G8/iRhrd-?1;z{ׄm_}vm L2;ֺZ. 7"\ei?^߷oC&V:"Z:^!?|l?" ~pmw6~/_{6NJndomwe/OķZMbYS?2޷,_RUvb_bطQszvǻ|3.%]|ܦ!xhDvڑ|m7fU0dpZoxH+)Qd1^#KrhHR`c') 0ː`0`RBJ&I?%ZDw0m 4T[%;r5l&)OY+G6֎KwfCslTg98JN"yB2p/&ÚsbǡBɵ@زP(4Zl=Yc- gZɳUٱJs w SwG S9rTDgv6B=v0 c+62.w]Cb}#:X7dm 3j2=:BtG}{{LwӑwI6[TErvXPbMDP+CitxaGF/|5D)4FǗ`pw+J!42 9|%P&RњeOʐ@:#Z$Ҁ*p &2M314f${'<gW%Ti:IW)%J) x:G+E@H_S8FC ACV(68Nv gH] p̅DSjKO#l>~/ƀ;#3^G)!WȈl@hK>'#5xPJNa8әٹ H#BH(V TXͦ`6 S5'e(E}vgna䵎L~!t+jCpIm 6 Y2#".]{6jؽ8 ?p8RZ66]:55BV[ vQR ʑ@kˊ^Zjj e7<`dmSTViR$ %P!Rܘ=xe vTv;pv ^9 P HiJѦ晚YtV=30W4-?{ SN0ņӯbeb9~޻6̚Pk~ܬɸQWޏ!X>V(g!a^ƿYRIe{ ρ<c]βyya^:Wx&O_fJu.ѩRw<âT"?̿b瞕ӧ` sZ2 NθI紐eJ~%Z'k] T~O7ɤCp]P!k_TncAWqMpCf;On,O:gOq[4Q;:<_SFu=PkW(}G=H)gt=`i2<֫3=!;eNSxD~`lvts@gzo p¥kRKj>+4mF_GZqzCR A ڏ:zZ>Ԣ^\Uw^IűGԭ8t!}幒+$:Us~J6S+t}==>##ؕSʛ,}ވ~)(G,oDƑfћMg+A3"<^2{ȭ ?a;l-KJM?pWGj+M$B C$ '$"*1QNdBH9# D1u&v1In%:%ZPS35Ja ~>56 W#x @cDT ΁ B6ʁ9V t@ ls0)6Z2eґBw6]RxxCTS`& Ht牲`I4#|̤+T 55$!n] :gb5Ob;h4>f!f2>3ܤp [v^B\8i<9X9xnQPj/iU(GsFs8w[$\k/Hn0/jH1vi ^\9q84SND|;w5'ԣ_.u3g #7#OQk$Jx')!1!- 2ڢ=k *(j^N@ƥ̙dcfnXΝ6ªRʏf14). ⣌WPsrM(wR^1xLKu9J:nxĶWEŜDPP(_T?zE|qA6}ǫXy:r\JɸW&p ]!`㼳õ :c+%a-`R\ ̸:0=dw!F Fz&{vo¦gĪ8.J^syʀ^A>q?̈́{eYqhviNBWD20rwg_k  ט]`$ m:LJWj6}x)X`i.PǥKudGmTf9 VA%ޥf^ JLk2i\\`)&-TH-~PBIE*4LjPVa+vicjtřĨE9Mi-%nDY:S$;S9l̚1S3.a %RA0F5 µ<,/xf(G9ދٵ20 ^YevU~S90L @ ?RJUؔ!FM/9rdveyjP|9׋xUGE"8%yljKIV @$#)1hRbv5WUS8U玜 ^d>"RiSI/{^75&SR~WhL-N &N=ا*!yFz֒h-7e)8TiXnNS&8-7;]4WsWCq*aagnnȽF;v6?-OyY)yG{*4osKu gdoc:zfmC^zaGYj_R؄Ě־;k߼MfV JOhJ͘wKԒ |TCZFd6ݾJ6Ugq95;=NM3+/ tTBu%V-ƚ0v?ߍQ8:.},DLHv-Lsѵda_5FRMPpQNz-":&K2at+xsFCF V!-裭&'-Pp-CS6}:<~3nt>O|x8 *v | 7aM8߷h#eqjmaI ,&yJ(7g >RsZ5RnŤ&-.uX|¨`fӥ)%x3D wr4Ly*/Bv;im8paciR(?Þ^7ͯ "媥9WU)*;8MwӛnIBl U"!lB5=z>4hlF07%ܭg_I#giReU+1Viun@ $·`R$+e~?2ꑑ%i##82[XnՖ(^l)^M2\U 4fe?xciLbRU=7MŅ`{ǔ aw7a0-CjBT&Dr8J( #N&1QFٻlW &T׾4'f218M[O(Tv_ 2-QsW˩r8Ije]r۰oW:<{O`ukv, 6TJ+X'1gEfoXѤ{]äk 7빑6~8 @E`A1h8zL&uԋ=]U/cw5m TTVO Ẍ$UqeN2+';1;H}wQRd=P#fQ@y,֕+@Pp%VPdIYLީ7Ia=ub}1"3 ɊIDrk bÏ>Y̷nT}kic(2hhKYN, 7t3P1+o@: y0dJ>^Ⱦ+fCv4YMik50-a %__C 4?hÏJn"%n .zyo1 MX$ #O0}U (&0IBRUS;0(ģIOH`J*2(\8o1OfQ{͂QdǂA#_T( /FNCׂ\Tf@wy>QR=- ǿi#[CU͋yݣ(ܣZբ)Yܜ^,ZLx;Q?c{|W 5GAw\1MEܬeὯ'KXC*\jeg2# ʡ^-[q-g[j:OYˡ^) `8Bf-嗿Z94O=B*Yq4i0Yz9>/Z!Hzt:8Fv{W Vs`c;zW= L臧!8B@3+wvxө |wVIbz6OF,p҅ 6A_G<'h"Gч;LdD;R8:HpOhem8agrIGgEBt<ϦSyuaEP^Wi/ybKRAΘGTp(*UNj9[Aлs'dki =#.6uq˸m9.WŻf8Ji,[`P5KIIڱy)i <`KIr9X(  Y wj~ZS0էZ4v !/4pEh4Qos-`:.8[KWV(>d/3ܺ<.6bPs܋;ēKPP?.v&n/՗ؑPC&3Gix;h$A^IBƚ{:XPiljyH,P(oL{.Xbk)?Mwс-^0S&gr+o'jrEmNY7WJJiqe>aA@֧1I9y (a|?Q 5 v#tGh.ǟGF96sh6yZu=Ѐ}r|oIe}OW딈rӝhRRݰf*G˕i'z٦Gt2eԍmq-cUG&a-xH-mŧ-t1frXQU~S+Zmﺖɥ}OYukzonMEA( i#p0 @1`$4v@M0( D"8`SI"HE*JP2`)P aKuڶZ4`bvM}fQYLN0Cy90[qfY>ŋk?S3Ko6}]@wv@,{-hoPi*k @y{t:/'`?? (=x)_~XA[c"67FG@Z+C̣_s%NJwrS$Nu6v@KG׫ށf9kE:HWH?<Էë$0/}6qNsty-jkw}‰ zHg #Mvf3@N(GuigDOruC#K͝r1/ ܽ}>0[;vMZc8,#7~WO1'O+2 čeD ;Ulͨujj/k mBn2;56#]hkWfm܌`&p)/[w)j$$;zh9^p'eaMap.:ݛXwէB$E|S4qYB eS' {g}&dܑٕ_JNfct#8ΒZ.6%G5 (^kϊ~ap: A(HGl7[G}dg3(b0_jܸS+!Ӗ"BԼH(_v8B59Bɐ\ӫgTbrX1MEB˘VUL1U= F>Z0; D}iu_}Zd P hU?(5YB|irS! Xޫ\'YpH 89p`+5}!0nȯ#ՄDMV{"y0,: զEEdmp8J8ftvR +[=G/Z0iE$Ms~t%$41@!BIU zRLMu4WW*4HA@( G$I(b$$(@.1 ap{9o3~o!'}HP6]}qZԫ oLgFKo9aMs0^sOǥ}LW]g5&Ybce+qZٖ,MրH쳥 c57n/Bg|u/=ƭ2Ϸ=?%]]2T Y%}sy`ggݟ2Im_YLkW+d,X9<@#-O'ٟͤY>ehMë尹 w9 ӌ.ϽY[C1W'*-\ PH6 bbǣʋۻ/UePmF;մ3b (#$kY{d'\o[뙶~O6LyC>EW ,D6q 8Ӵ9z!VOKv`Egg<I(B(*"@fUQ՞_ G^#I%z.ҕHFg=lpӞ !b8! AH!XEI &H) c$S'4@u%}8TMw+ܦ2:֒JYh2ZS]1*1EHqypFcALd!Q2 !BUyT>4R h?ac8wzZ1) @1^R/%B B_=IQ,"7\3JzZJM5>R0=`0|>3G cAZh&qOS>;Rei#/)dn.?Qϳ T 0[GR}C8+s@IMA3 km@oZ󯹺擭-РZjW+>#߼}Nͽ`6KFCY$&|<}.o@hl~0PP4X. y%OvLmlC\2ӯ [8(ʶ"V\''@qZ'o+t(F铈؇헄 1 1)D'2_$D8U_Z8"BD*/M4Crc$7`ݯ9]MjVz+ "E!(*aX,euji&LFG]Tga@)kAek sb ޶-)(οR}8P1b̄Y%L#eͬͻh,t31kek$A_p1  ffߛ}of}qf/_MӱPBpI!{8zrS?w.k.2QYJlenbjB^U6`GU.փfGMa},(Ta bgl*,1#@#yP!7pX8$A !# ,F|ۏ]K"ZE`v'Hvȶfᦠ)ƲN S;KF|`= Bh#rv}5z%ݔ>eX[um?̯Q-sj=/n{CIvk'5ASQOp^ض?2fZq8uWt0VqεTZETr Nu=v,pm- :X fa\"xz `E!ESk#[s}HDKQ(Wlk"_ᶅ0w;B"B`2V◅X®ֱgGVsX0!"u7#!}Asb.+9[bk8¸ R%]o%8N&σ"D ' }ip {)ZfVWg[^ю.b;lNj#!@ޓqdW}Au!=v6dO^u\ I%"VuW_QfgszLd|?!#9Uo&4DV`-},OK0Hu["9*]cx[%B"ô7` ҌQKg k/%#lM*mI@w*JsKܕ#IDBaA9%M棢 )1ͳMowS&՝مׅN]A K̝11ŤcJx+%htr syNu b,C>Rǽ z@15@;&HOL4t mR+Ak :Db4F'1YMRBJ+uEѠc$"pMtw# ? ğQK)s nD *Ţp@,@z (t Z"R.5(J8W tN#`7pYG5CbtJ0n88X*ƀ H L%l:k%(#?!`dO-׳Om.ܘfhx2;. =?b qh.c> !iNj_Z%: ` / _Tl%hxM KT>-wtx!iw Zv^Ch< Pe5 g!pp`!=R.8-謗i赯/6xc0R1*Et(nqTTVMYдyrx**uk9 o9 ­2[sr9.#YƳxD)zngE:'#Yk\N̶Y^ !)NAGjW )2:& hh0 PgD!F98> pvz鬨~BƷ ->VdFk]zhcCr{!_/֋Yk1iub ܕpwW\wo]wf]{oubponQF6^=:Sw+JDGVRRJ#6ؾJbfqWjηsl&% ='!wyH;ۏ~pN{EܣF INԬ,29b99,LSo6_idgYOٞdgQHfE/ci2fK7UGI JUNDK ڹ5{!V;2[yʺxJ;rvw3۫G?s?Szo:۫G?\057W\\N\d࿨Y%sN?^}'N%V3<&K'[$3qɸG=5uy)p#JOhPNKd) EXQdcnknn֝FSKbqO7Oۀդu-kl/hZY׎0%aR,^Y&,)xp^\YBC GiBX0bfWD('Ũ#Mr2 P E~cy;t~UBi(|/Y޽BE Fߓ,=$Ť| DQJG TU1} %@Ɍq+# 5&KZnI_)҃.ZY(5> ީΒNuTYBB=]BoL UZ  *s @]%DXFy 95h6Mʺ^4J2QbjCP$ q4:ɔfZs"J9<(<9ob a@ KOnƳ)b؋RQ1c. E@.K&AIlv+A>v,*rIGJKB.wCEL1Dۥ6I^byqsy}QZ>x?Ҭ?:~Y<]ؗ׏~d-~y)Sg ;l`!y#c|zA¾ZM3PƭVPgl#tH-~wVbɐ8Z\5v}ELwwb(` c}RXK(`IiSғ$I-`'8+=EzZZY@zߺ1edYg}- D/UJJX/{uGKj3`I5[|MU. a|̡d6J* S5eҋ crrҫ gUzUaMzeI.q''W(~y|n>'gֶtYŬw?xXz{5piX~Hs.ί7Ӷ89ē_r~)|:q)n%C[?ޝ/~ J.ޮ^,.^//s^G(UX 2+ hF}M"SL_e Uǧ \^S|(Y΀8D|33~AXd?8+7uˍg!@x6%E66[͵,6kkwhuYCs& Åg͔CyD-q5RGIҟ|= .3E.XKRmXc k K[#$ ˢ⁻@VrM-no3}+!"ņake e.pDL2U "Iph(aܣ(k4LM7͗W-)z#}eJdb;F6凔;w_>ȧmfn2ێ뫏 p?@Jo77?]56_@]5 H_G(.tq~~?GSZmqX}5yu~'CP n}tꯌ &$"vʾ-1IU NK}5F$\rB%_VAPmhAH5dImV8S>j; k% +@: <-D8^`Zs܀EQ>>$džY,jxϨZoˆYiN|g #Vrԯwͤ-pb6Rq\sE3>}Ji%4K9Ij#O2!(F5۰mW6~7?(^g6n6dkv>a4(2"LWVSe,\PCm`V"M!f$#qεiP pCИ_q"~͢^n(&%$$M &~F(Gvx$V18 gsI ?{q_!%CwW ; }8'kviS#){=$% )rᐒF=G+8.cIhVk!iӂVFK9ryc:~fsÈ52ن.ӗAl0;(o8~2lp/la2pБ k)Z/#a)oh|3Z$?dlC^/nA] 3I&ervcZ6_^.z"kIv/bɾ/FPQ0%'K|D[%'6蹯ǨC[}u~"=+v&](8>>N4x,4A16tuă.ۤ 52"J3Z=!]fAې&QMh64&a;HtrΈn1)&S qWz|%?^-?raI klRB֕01 >uG.[8u)yϳ.}h ؋hf ms9[_3yZe_鬓Uk/(%0(_F'zL]1/W .^''jȬtPSA*̐ex9k\r ٽh^fѼ̢yE) Yj d^9Jjk}.YY^DLgȺSwf3ϣ4zR$QDpkHQ"a}Z!GUKILce{/J6a%6]EW^p!F c&Ǔ]"ZB! sr[{2Û8;VM-H5pm4{?DoC@.)UFش\DL!pA: Lk>:]2@n]KQ`DFټ@*e:iB&*S32"F1XYǡX&ɛ jOV23Jӕbd ih-4S,F9g28Dpd b6LQ <GùKV3bԆ 51Lň 6QZל+%t6`O@܊Pd5@w6b?h[UEa ˝r"z h<r}JY*).KID,Db&57XU F4gHrn)Q09d-n.QPsZ(Jr2 gaK̐ki4Y,@CU\( |ZO`lW`iiM+cCY[(vvhCPoLKk͍Z'6JLN{N*hxg"-e { !#mmDGEJ9CVcQ%  90$H^=CTVd! 5 ZvP`AO `A;45O#3K`棴&ϻ\_W"`GnZShBɃ6V:Y{qHnHB*,NxSׁ6Aڕ=K6a͙vgd }EJ֓(R2Z.Ue} y '!}@'keڔΜ+--ggY}"`HeGcLpHq0ܶTO]Qx׹HW6vvؗƳ@)E%F|( vUV)/EinQ>?FË3 ޸mgr σzޛ=ξw*l[57ݓ%Pߦx,Cz7I]nf0f o_,Ω: rB*yz-^kܶ+5)m<ԔpR-)z0 -m.>giB.cn%ί\ OB}KW7X<פیKÕ'QyDEn|sOF+4ųggۋs._~/dzA`nO\ܫ?C=;w̋9S7w]̍ioS A(6syg֖']fۊjSyY"M<W~ɚyR(s8ߨXe KļxM 8,,^kapzskV)A!ϖT}n>8tyuyCQqVDfkد` ^'/DFތENLG;Qeg(q9Χ;[L;qj{N nE롑"BP>&쏋\ֆ9X+u$`)ƚ9\˕%ȴf"iq^%rfjy 6^1@:sL ת wϳp3׀厗ͳJsÄl^3ژ|!Vb5%AgX w5^B^HrvT˿ҝڒ^˽wղJ>WNQɸ4oOg9Fb5scd^S{L.X2wPt$jCLȰ,o _kz5(@녔T]nOJ < J" f 6 ApX !WruOFĝM;!'0m|Z%(m-X쎄̪jGqUA?;_t*OnnfUp0u_k"uӇ,S{2Z{WH=ue~qԯ3[AڠdDPWʐyD%UmS}{oNv%؉߷ Țo9h')K{aqn1(WN?ĀkXز947<^>/|^fyٔv'}C7 Ԝ=(ԻՀĴ]ޑ٫7GQ 3ۦ!; "cq;G_0k)r Y)0) !ܕE@lЖ)Jaɹ.GrFYf`GQA0&>Sѯxe.a{rw+4ŀ6PxMz*ݦ`f.)^h.{߆?]_ñ ݇3rbſ=QUQ5d[x~Ku+1-.^rZI uT8~<qĽNS;ގwϥC%*f_MY6 ھ-0/a]M<Ό9lY]~9BէOx!@nN_օlk}Ls0ӹbUj\An7ۏgX>8I<'.t$*% NP"r`M,}|wvG|LΗ^/,E)쩾%QSH_mXX\ fԧ SsOYى!W[Ѹ4PBPR~ʘXqU?W,ס&m}mе"<(11z cbПӃݑQmn1/2x}edM; UKuEe4Ǜ,jތ6%_v=>CF.hƹT)FKbkXRo>?-%5PXdKϲqr%mrGE&wS n,k.Ftȑ-(ZOf'sJW*oox%=E7=k=]\r49]m닿I~ R3Fx<_YdX>Bxg[]-JJ-aEWPC&k<9"YEHO:yݙ'\>Rį?х1i1s= mO:ڈN'b/7nK#/A9_SwX`gg=y!o;v帻1Ue[(QTnNRQ\H#^4- E˛ÐX)"4PXd`J/%/ ݃j\!dYaxZ6r2IaXrDlfQıDj֡{QҘ]M֧Kxj4un} #eΦ1O3+g T 2fc#gۿ1dD &pzi0A8*2 )cYQohyCtJHL j:B0\;^svۯbZ[U\H.],;K /ܷ˞' ?ן@6ĒRGئ@K,+tZnq"gFXlilr%x;驃7Ohs7J܍Q}I.,[)m贶wgU2}nsI{B)yi{LƷO@TΐkDVx㓙GNv~6ϏMDGWƯ| yhd??i㫨6w-;sq&\jX1 e4H#=q\iw_]|!\s_.ؑ! ƿ>@k ьǓ#uH]EGH3LB ye'4x!+/Gbi*EM(msC|uU |O/uIB"]+R7֕{;R :ۑ((,(N+ޣWz6 8af&N[_o?|<ԃfX`je ĨZmB oN1tgL$o3&Fɉ9Ð09Jc:;6k,G@N-ETln_L:| *mj*O(埚RHkk!y*gg\L& 8 &3,"\l]vZ,)l@C@Zo&!a66`mF\P"C(D3Y %j!x)=C|W2(vGrhx=Pz42]ζ3/a⽪ݜ/5_=Vk+/w?t*Ԧm>Uk2L4[K0_#EYbv%nzE M Ƨ)] 'Xg8O:4BdmqoCk,mHvD==4<}ɨ-B+e*]"^~PDBFqׇ qt}꺞yFRS4=}mm{q ?7:y?N\.S 07;sۉ$q" 9Bҿ;ʭt=6:k{KrO#V{$Ҧ(=ۄ7fGdwOAT=BTG'S>L*b%o\6?=:83*3 hf';& OO! 8Fa9wgGW OLjxzQ N^Ū1gܛ1'@_.Iw'C"cHd(-e i^.6YJ%-Pc\YN!RRONl| zEJB<>M,ogf|K)>D Z.w&e4MpJ]>Z=6%^\I n:αV_@ZZS!HE1cQ*  /_Ƌ`$ωb:ڀ>ɢ=ZLuX{Y=~ pkNi7Y8rSOxaohF;CnD!J)&0Zɸ`M!Fd:`1:~Nc,C ZwﰁrEnŹCxa-osIdeFS٭qrV'>d87\܉Bq@.jT)LV4_*rzZ˥G䜖ci8K"Zpa*Jd.zteG9ʻV Q{E59tRDʀI+eEs0ZRK0)"ooޚ ,Hو ! ꄍMA h `ʲ(,UP*/@Ez Fp( T[eCAP!{@h B%4:f BF84S1Go᐀y+EDOJx`6x 1Ei5IƐs.h J!p,b4 'M8TF8LH[xhEQf9D%WT f8 X+%LcH(S%Ti(țlKvR-9N\ Ls)FNRESțZk g& rG˰J 0$|.YZTb7 eeYq)Z~G};j.n7&yueyU:0q"V?m(20_r-&NF)h_&;܄1(GwaM20c UljCʒ@3ulۜAwҦmנ;mu[kS>[Խyt![@w!ڌ&y6Ւ<+[_wκ $hLq0ЌS8\@J]v)/Ƅ,//py1f`y1`y1`y1̃y1>`^58/nv,/nAuhˋI9A]/&rvy1i#ˋˋVˋ)ˋ):/FTc <hx@&GI?tj: IML89X^LQPSSM3D33sV׽Œ%A.%ːJ-mOdQLF6ӷb>]\8q)_[e#dȪx][YH6q]jpʆU;"LP;`6=]jN(QlbԒSqaKIwZ3)w:L,5H]ɥKM4tmr RkEޘMvB;ZH6i]jJ֏3ޓZV22a,L+)Φl9Y,폻]iʥ_qr)‡tytIlT{M/'f@=V'i—|hx0"5 I0!SY,@g V`¢25c!nY6+C#߿(Pr1c:&,5[r&eSn\~tͥ[n1{l[?Ic*ޢ[~ftkBDlJG&w/1cn1{l[gMc;%jF&,M4Ǧ~~t? A1[q1iL9w7,׎^Cѐwn)}@7Ш'0κw915F=A՘D טs6%;W2Hϧ!Ԙ\!e_ck̍z"s5fט)\a*s_cn&13Z_c>3ם1315&=]w7l~}YO`ü$15F=Aǽ!Ԙe֘Fs_cnݫ1íI_ck̍z`{5fZ 15&+@(@??_\=H>CĂ2[1twU-X%4b(GLZcBx"4ƔabDxk- $fu eAPye rxY"|)L͒ǃa~ E ,I9^rb2~uD1`tpb&-"x{-%$R0灗 !$&53epT! {XTk>M/a^θ"!"47-+'~<+|W.UxCS1Qc{*Tչރ EEoǨ rĀ&!񨢐AI!iZ#8a o`3vXVX d!e D a&jJ ; 1D#l SrZ#H2!o %D4ʅG} IFx#]TY ө&",ɂzw!_bpp$ #Q.({1|ޟNXM!GY68i4rVkls!JyP`Q<؈Gq EM cDHDR {5D0X/v4Kp)z3"mHt~7uNzE\[^]EgqCQZUzeӻ&/gf1Nv}2ty2Ǫ 1\:|g\zwk~[N&p~ʑHosIlUk |J+ ޱFY#D1o[ !H+`$F8 iG->+VW4a4-s۴ <-WWjAiDMJ~PYc󙙟POW1>,:X4/faJ|wY:+Ouo֤\𚼕oaP[HM,Hia;$o0xаnZUjHRq*dYZXq; 4ZY/\@.m^Cq/o5k\0C 4Ր>7œoHJ 9]nƥUOb:t; {0IB>Ȑ6=ev_/s3v g& xVѻ )&s1{hbw%Bn]BQ<7<L|̗'<[û*Go:3rEХMϦE*DvMp+Io 4e? ĸH~k=7{} XWl7;sd=w Ԋǟ~ n0ǯ)Ϧ4,liZP4^yiR)2 - ͖oMdԧ텚.$vB:|}$>^FKBE"|%U!cKyՒ!*Kc+#!ajOeE#HOzUօޡOfсTҞt׃( )HaRnaQO Q'Q`:.%4 赔o.#wIcW:DJETDr $ܿZ詪[5ʿ-guU!<f _*5@GF)/_%輔oWx]hzS.hdx|8OF3f:F55 aګ/ه->{0=_~{Rxf9,TЊ'.&Qhj+NHR9yug  czx:M0t]=>l%7(x DlƂ 7ü%ϘzHKG^ҜM :ŒwX.zˬ ׌ R3M.5=Ȉ,cH~HKhf0JP9X^?^;[E^/=8*nxTjEȭA&\)HF)^S.9 M4{Wp1*$F=@1:F1Nct(&<5-m#FB)#FW\jMG_ 'Y@Ƃ![G!6)(-҇{oW\Cۓd֣<-]].n)c K\z;[fMlLF(Ly6~OPX>O}]$RleUL_V  JDmdXkeHA%NJ8g5A 2Si\ݹ4△1+QV`t)(YR \^8ɓRIy}s.1 Gѥ2]|8r4hOB%f,2E_ KEy>bE֍@6%.G[ Fmf;oOAoʩoX~?IQ9L"(sOY ;+UrN}p.h4 eGT4iWYР!;7]~C$IWaqJڞN@ۧmȹ<84{{'Z˧'IgB ٚV0D/!_ WM7{=-C ՔtUx$TڸKShSsrByh`֢•aQeoԡ] .r}jGJL o MbpDt5S{ΠGjWRA1D]kI:A!"tFTC`%OB \m*>]TJU ѾXjh){CD7#N/ 1LmُϣKY׍~ Wԯ;^x2*x:<4c9#G܃/$f$ <0pRx^ }'4wBjƐ$NHy)S` ' 0='\On>D!ެyG~v>f) "BqJl䋧!OmYk3A\HEú8* 9g9_K3 y 6& QԈIa+==Ldhe~\H]Yao=ڴS@Y- c:n^ھǣ/^#HwWuW(P2I %һ+|&ſ!7ſgfvnz7]~*ؐR@0J=KEVl1hcP;A ԴaF$S$O8mH%ao{ %,UZA$>z= 2%h bJLZ=ADf^-S>)_%, Uu6FoI5X:ujjU)I]2%+E&#yi ' Ӏ``Kq6P6!a@՟jJ<uNZWU7)ZWUe1&;*lXB<& 9:&ިU=\*IUCdG hAël,I,&bR,Q͔Q <.#(E2fr'׹5AjFoF'GGI>QCG2#HDqӋġM5Iᥓz$j p\8xjS":M @3gLF|dSlI&}SZ+.@6؈JȎKJ$ژ@ (P9ŀqvJԲھ (miF]O-n֤1=D )C}#:Q s58f9d)w$wcHY68~( R1¿(>f,ULsJJ %^9VB˘)+*TT/W_-EdAb(iWYV ngbj ݭ>:Zx%PVN;GR H=߿_L \QEړJh{]l{nkqƲEЗEf|? `0'3_2h,ҭN%(Z,e[bW\^.[5yT㻸>4#ޖ_09 U-KR=$vIB{){? D|_x+s?o{Pzs3m]v4=ςf]E5 DZO(84C@Œ`)ʯ }U6"rJ(5!*%X}{@_ qYWqU^!I5 Of4Ou{{cfGp@| 0 ;x+әG O]+Lq̕*$!ιI"dP#Fxwz2rP,67A4ah(/bB[z"aD۶_d\FVd)Yw.NO^^oz!yn?LYdMSt)n ȂUAOOە[A}'R8FOQѳ/Dd8^,"wzAwKvNn"47Gn뜿[׌5  gY%[n)yPҷ6Ys8{*8Ѵ fnJY=Zh㠜(UX8mu}h*.B$Nk0A͏jOF! 8ύPl ׾]!b)ͥ dqj:n6^nһ&q֓3yc%N>(g隆O6"8HbathdoW"߼郝l6hPu&+׈=U0%{XXP4ԎxH}R`:C7rMgMp}3_sQ:?|7avo\[G,k}>=Rm..俶Z%/C/He0Ed4XDFNdޱ<KrBSZD=XDMd9ˇ88eLtGn|-b#/.շµW*v\0`u{/x;a"W\bHS o%APٵl\t'{S-ݫo7Sq==F;&T:Y lN9 67FBPw5 '\Q=D2)@w .^*k eb8SU"O˶ŭbiu=NwĔTwx߁^nglϣ4^HV;c[j40yhmxx>4fM`KzUVVrqBey! Mt LW֔PВubR!E nK- Wa+{ݱ)mA_C2?? NQIT3Y""(m vcy<+^L4` H EF,rz\B[eIz`"EW-z>(ʡ7-r:HɜJgk!RbrFBY礊>H}Q@O6eߍP!0c',$BK$9c1 abp)Z&oD(m'C{>wvS{aX0zs:)یT&)YTh-(,wYK)XR@saCѾsd̷D"MBxBam"(PY;sg3tqu:|9C\\pxKrPbʬ@o8څ,PBC!c̻-e^谵 ~56mcHӁ,-6:@@k5X$l8m^Ym111$ikiZ (^$A%tB A9g8P*,of0.4 TH jhr Iw[zyYs W)0n܃_vW;A6Y1iQMzM-O`l<Y_өFkHFB(BC 4SWZ+%SiHZ%Fr,D}BW PKZ+Ommm%.q׾#B6Uz?OJ|?pU6AjqYsQ?3—|sЀSzv}I1+Ƴqh}'w}Ϊ7O?p:/" Fґ}۽ǻ? T"O #5/*_x2捕{Cz̸xOOQYHao1~q}3{O١c8TD[S6+_[!u~zj4S#tJ ~n6[7G 2MŢVӵ J6DwKO?p.POl>\s(r J`gϣy-!RJ)u:&NZ"#?B)-s̝095#"EI/R_e*4ee Kf1RQJr_qcU i!4J(h㘃R3!_+G4޹dVYAuR=&1a +3x% 5"Ȳ},jRpQMɱ"OV!ŬrU{F'y9ĈIe@:QXDX;DJ'Z /x 9s#^ ֍[`( 5;/aS.! Qph&0Z#b~ =)gDqh5O2OS:Ljii @_iE"$گI'xp&]U.VXH*ЖBc*sI Rq<@ &tW! 佧BҤ[Jgb͠$4ĕYb/!bkhHO2F&][b(c޹R)cHQ-lPQ}mS\" @(!1@(mu۷Wh(Y2xCh  :~[I7Qk>Δ:.p 0# @y:v~BH;KVsڗDleEaj?#K3W7D[ HJZsGg)Kwj4`i#iQZ5%C ! [ӵ}\|dd|]Ӡ#wm=n; Ud] egAؼ¨km.7 t?4BᩭhKX uM'iwvBzz-V.o\DȔ[ʹ}jԈNwtnJv<{m ELگv-ڭ/5])c8v=Z6|"!S"LX}To+ͫ%b|tLz毞?y=O~Z UwoL^T.(QjԨ_=Đc53*M9Zs{\{&~6 `TB ?_| &5A)x(\L>mֲ85[V瑽L>,iNŒd^,,r J:z631J| $! m=V@J5]MXֵܻ~AKJaLh`[LG"?0,'!'Ag wd;V_B[*SAHbG£x$_TtAL}ih2C9f+ SȆ$x=OF&kU9xBdIrTƆ )p"'!XfmT1*q)fp|*4#[0#!&d1Yt[ŕ@cR)p6# |ͮ྆]S{ù)GX "NVۣ$ 맢'-^DԩeIDõO 2)}Q#8xQK" GfZȯar5_>[6Ap(yu}[~Irwv}&G#N맢;$*3Zt [;\vJ |"nPYLh92E։O1?a8MJUNt9#'/"bu8nI8˳s?y햬cqz)]jUvw u+ & ,;uރ` ڎwmEetġֹ{"[5{J BCm*Wvv}[$K6H' `7i.DO4KjƎw;K7w.Y}|>PLPa"#L$w^ih,< xVp] $bv^J |NehUa;΄S-ev` is#6lb·}m韏}, -pL9JBk͘j^eFeK#%ơ{tX=e}{U+e׷0ڕ%GYEGSÞǒ.2r&?H0nw Z#O)AfF8>x^xlEv%8A}z-05frAnt=a"OI>otiE:X3߂ (߃|qhd2s#)rܾhqX/:ЖuX]q$VNn"'̓Ӿ@YJ͘2]p@0 ˧{q60a^nծυ% hnG8 FK GֱH>{X"łaMUvXasLGn/y2kn7B|{-hb 7u˥͛A]}zvwCYZuD76^IjD?B{ݣMe?,^5n4byQߛ a4}<IWo&~Ջ} fy#lxm f xU?.٫BhDo[imVFH}j-Я6^Z.Z)\(8chKp0hZ A36ԾG2#9yۨ:pP;'pNLoc3VXq30mUi?FIf"^ImG U^e W;j+(NզxձƗ|m}25jgS;gm9m.94;BYl]=]TQ1@4Hnihc} (L)v(_ ?"U]WAւۦ^L*9|vj󒋩 9Ht117@l~,KsG~✡|ߐB.|K c$29z1JO G`|zX/wcnh^d=C4Gk+38Rpˍi]rEޔa*I5Sƥdrn=a* ƺ݊dgH6ѳLvd6T%)<;:#Ǔu'Xw}V\3*z贎mɘHpX` rR1z{Ř9P[WO vh|DnRЛȜYeyN -;=Tϼ2QKC `NEH̐D:)VTjۿ@7rCkJnC*҇kkcr^gHɆ*ECQ J@[gv"B!@?Z5̪**0 蜌JEYWe]R;yBf |F5Wѓޗfˢ%l|z9rVi*o$j-nʀ:@H~OQ2LO߿?z{Z~M{UF ׳wsC'-4{i\]3Hl\0a{7~;DR_R|ȸF sc ΦCx.-'3A[[ 8W~\xD7%P˝/_e?zƮv7=:KjzsҧN;Rߤ_L+&,L|Y0en3!3AjӉΓV'2Ag< ;`WD؞>[ɛhC«/zvKc ^-w55o#)n޼j*^z:~,d16 kPUԡȸNV9 24$Âd ks !rW l¬ҁ "[PhgY_)Č:8+Fc(g!Dd&G b,D`c* O%%-"bå뇙rTs{LJ-mc`mz ƀ! 2ERv,]¸=&GMlT5E1 ]vRyfe΂R#V:zeyJ;hjKOtS2ꉃe;Mb SCcXcQ4AIҲ%1t[G_|d0oشd9`'yQ&ɍ` IfTsI`Tyz骾 ۦ/UgV|t ƭVz1 3z*߶v/6,Uu5%*ArjA?T%@!^ XU<_%:N)_x 1aF@)$ 9gȝh_z& Q-/Djd j|ZP-ʼX-2Y(0A'%0/V>3-Dp(f _˗?Ϳ^v$"H-7mLvOoƭQ/~?urz<}\h??|9JhOjtW(,NNm6W4u!p@?KoR6I_M#2|qx(7*Os;wd?;?^'F]o&H({uC^P~f`zz\`u80yo_"Xfe4>;G+PS.=D օZ෋wѭ~mBܭ/zyxGTz^߻.?tMbn+;ǔDBkS.j<$?u923Kğ8]M"L?M\b4_.nֿr/i+xqu/&jUB.j4 RWGH' AfjAz}aS7ݜGʎFMZgSث"ZdVer+;R֠_q=mgRs`89:MJ.gu:M)$ELKmS8x|K}{즋[&QiHpCdTy& N1gs#/m'QL'b=6Vq2ŶCZ P@1ݧV7L'@QLhc JkA/š"npܾM7[Q>nmX~#uH&2Fͧ{nt1ZQIU%uL~w֭?`u! ^/Cp4o-kmJ芠+{1;ٛvq%g",7xοڿ)-6(Aoz#=_ >SWj݌g2Zg~:@|]&7mbPlCfLj/PPq dN,/م6Z>G4s:g C lYJuؚcS.=R}{|JowsO2i}32&@GʵpӃ5M=ú: Oa{)-m!Lw1<(e:1v#̷$)TvO0p(mrI>3ǹcy SILI5!MWi@bDJSq95?VL}gr\ie εy‡?f3ud"6˴IGALNUfzueZ͑aT-"^.3>`;Kl Tex$ӈ;ha6UnLNݞ7@2G|$_I?A V# O9/LhcgU)I V9d5GZGW*3eSQ\jbZ!O?Xf"BYf mllϑ(3*6M |6BĆM݀uZLZiSͰaJ{btNNODplwO V`3jrAy]2a=wk;.O0ܣ+bs2sZ 癴 U@=h 1=8Ao$`E{ȃ=]F txr|$gBbkݧ6Лkj%?uV%ua;v; `>5ju$ `,34ơ5m<^j 󽺥'>c2` [: Sز()R$; ,*yuzn% #`JuM3ZPcyRhu=Ky=154z5Ai<@cs/ Ӷz9䀘3' ,h.}^$S kS(0VdTq yqvxqcp*$H@磘{kxsgs%8 5E e`=m,U7qDNZ=`2D 8j=t攜1 U*O:ઌ7 ,JH^Nf:\dxBnL'!PϸQp yl@ {ur!:jm2dcX!!|rF_eE>…=E(qqTLg"c?EЁSqΪ/m#^3w+Ii≷)T!Hf,spI@Р,y(綍GtUi}Poڪ:t*z97{Ƒ ?Mq>Ra[g)k,E"l W=$ћp5 ÙW]N e@Ʌw_oG:cp4egdр <.qQ!K^qIYZ7M7s(@ UpwV;eQ:J 5*ݢLcַѼh,ݨ50UթKsb,K13 b3ť"= ^G3,XvN*.;;VsYpWmdy6ϣ6 SRqrLWhgM\hY< _TT<% ˞ߖkf㺂a\}in\ajWh]q]}회kU*z%+ -T,8TӡUk+_Lr~Ѝ~:% M5L[U.es@݅He#Êe UdX3]hZ^f>IS}?54.OȝX,/gvӳη?<6j,5$ukfO@;+sDx'x;QўNt4M_A`XYn00=EbcAHcy/XI`JPʑS:.iOjF.p\‘ejf|2j1Ēzo5P6[Ph;>hu}Ɨ{x%g"Ă8S,)K%Kʒe˲q0X™я)jf^L:$XdzPI's*4'\Rk~*B=ׅ,'[Ɣ-KƫfgaBBV3p_=V֊9ϼ=Y"`OcwSvEi ALnbt="'\ n gT&ISux-(VX>(}c;1ڹU&aC%X Eirzm^z-O^QƲ 2m|Rj2?pW<>E_bW()_lQ#EXf\* ƛ@ pnS+8 sYg]wJim|47 >':4!#׍h -Ti:b#ʒ""R= D 1$ \,QRq I#`cKF(27*P),T񥵜]')LqMT񠍡{%;4,& ^ )tA7YߏoX`1/jkNj\=cVHQ3W"V5h֎O6/sw-+EŲBRQAWdYM #/*Y3}c_ʾ-{NG:@0)opg RBz a7E.Q2G_J$9w[<|XX+^( цŀXʈ;L1NЊFL}T2걵F%+rѤ"\kp %,/LX"lye*.A'#DMUC|J1t'H3E0E)B E .L1C)a( `x$?vmvm|HI _,WgNP΄@̬KHaiBpP IBhD~sH|R~e;}|IhC5 LHbЙދhP냓Dtt.xd#Xs6_'dzi>l~Ȳ^0i+ZYvLo|dCf2Fmf=?#Ї!Vwo5)\6?9^ol76Цh6\4rrO}|&v2j> tym.x:? f[]- {dz{;6`%T n ;S/"y;y3ݣf 駉s _Fޗxx6w OѳfCxUzbsOM_0kKlcPw= -τ.gɬ?͇a<šg}ퟦ^~}Z}~u07` Τy巜/w1aWv_N{z^ Q}'+`};ɥ#ងyNO~WCL]9߆/AUru apSp.~zYj މoa!$:3څnȇ[`=?~כ[)`Oliuq/;y͠n0}fg/ї8L좁3dj889o5r68{ĺ1Sξ8>~1xTv:l0j"0Qq+rX|?Oxb._go(Lޗ2,? vynS[0wvow>Ol@k痸lʀA\8v`06&DU`}IFYl 783Rpٙ 6h)Q4ؠ 6hpuhB*LgeGGv8HdRibƂ-2kZ#%e(JШsx*1^[@)QE R(e mswA!";.xySs\P94^屸c>L~Cwa势%;\:30?5q/?5 +ߚaBhc7O!UXU SU%ЄgU_n -W쾝Tl: #ӚA' ,[DhxR{{> }'݈z ԣDzexD(I(O]\H#/)>b؎H!΃HO;g9b*s)i.b!t9<бrZ9,rl|`#XX|uV?> A;T烙NggggXAHxz;g`˕YKl)tzh E.nVH>RÈ.ztcCe˳xth-ZnEoDbZ.~HD‘ɦ!JC) VX|eXJ/Qv8 B`68&!<{egXhaiPr&U`%,PXPKB4BH"6Պ>|p'cd"$)&&DY _ uȽcZ[4pQ Qγyd*;Z"Z(T6i@(x`c:tH|^Jژy(#4V c6"(\ʠA5A0"N: }A@_}<"4I2J8DY9p4BH D9נ&c@Au0R/_D_[¶;5+/? LZӜu &%\̱@A5)0!iQu"XX.DhD~Nq,-7E2WP(6ThikQT$/;erPhm]-TJ 3as/!Q 29)rn<8KU,@2X] fA$ EQjcrhyj-|1kF2kG*70A1+FL(,܎/'ڀZ;h (VTǥӵQ1!r7+RI*K7IsgOL~vϧ_*>Nce*zS4XOXmڦ?yXI39qLF;$YÚE krD8)ңbY<[[XG6^H/[b&:hue3pxgkUZFܔ2ITRb(]ЎPbjutps-Xg,c`*K1a)F5.Rkjxg3⣯li }`#ˑWR08ڂ) bpBxki1LL Wr0EA)k &ƛZ+]kdfFf6^#_2fR8 $oeKf5ZFSp ݩ<Aub('cxeRAN`a4 $'O1/%125qSfUg G^uUSx?iyѱ]zG{r$1qVV؟0itXFF~)n}ذ7f{̧pwEÂӾ5\Z F$'<4~r\3kJ`A[PS[oB˜غ쪵hxE˼ -΄d߷-M]VVF*L*uusUT8FuQcg%5mnT̸,ULQ2tWȦ>|] $|(4mKky1bM X9>ؼ@2vqA 9"L#8Džtӱ٪yuyfD^3x)AZ1Wd&FZO&p.Y2 OXMP>h݊/We\ 5adBf`"њDk$Zk֑j`4X;:%.' g2dcB1u~B1y˹>C [F.bU!ByMr >\Rhs 4p4UTP.T4{AAU!N)}RW c'˟n1SҞ]_ %E Ea U' [;;!NkpZu`\t!6 a"7w6r  Ȕ pk8'$ 8 L B8]fGn3mG}wY^7VEޓTfcx"YNsȁ6nhcdo$M.δM酅Dc.k@Ku^;?9ѺV}_""(ɔ}-˘%\Qiey#ƛJ-E׵LY髒{Wd\|Jɉ+= \ XVQ5RR\=v)[4O`k 8\ްE}3TPx"s\iL##Ri%f˖T *٪B%OrT(.gh{fc\)RZuOb-pG䓅)S{UێŢ#{[uh+ "FWwfktCG2而9FRyscj}L7Fhi?Typ$UEd\Q|H4ZW~qy "6Pw㯗;7ʲ7#0jYV۲ Veƹ_'FFbG O8>eklwA}Ixf0?hͭ[ô<|?\ntZw4vv߼k!w?G̓7{;7/4mm|owielm7_ı_Y f/v[~mQࠞ3v^ڎ 77OIwBI1k[;O^l QθiĹ1j&ݸ?"ߗu|r`^}fz| >6^~1 C9mYO769N.qlk)Lpz򈩻p3TX5^.MazƋ[6|Q |5.P-_Ac~ӈ說]S/%*8 efneo».LyQn ґA0S7.ͩɇ3/Ư.؞7w@z^ c Yj:gNN:[A"u;nA?× Unw^6[.Bo랝a$O0mݯ`m !eNOz}<j !id_IԚqC㣆REl)ĥP c&xE2:si e!Jouj^2 g21lDu9'UTWW@qKGC$͇`[s-VuT9)oq!4YG pZyNo uZncaysstHi/B)ҁcTy%IBTXI\(cX%DM_Cu~{zQ;BYP_II"Slr!|V!r`2cg'ms~ƕRz13̧8|rpbLN^|v/_ 2G$ n^1U㘁/iϚOW0l/{ ʝpw~$6/bxR͵A`㭈$"| 9ep2[C.^/(i%@c.K-N4Y' WB'ޅJW0an !&r Oqij{5Qwc))o/C':9h$y% ¸hQțLy:_,L\h/U]ltZqP󡸌!&f2&v6VɅetǘ&}9c'=z t$Ld'|• ^pJծzT]Ӯɣ9j{7W~tvw{ZrW`QٲwBێg"V瓯%阍9rńX AS(}&tA:+B}ħ`mc+v"}b}b++nRJ I=n( "u$JB- l.*>bbh-u>1o߾KQxnHtvɱHg&4!ZpVhfnk8p:7̚<37G3spd?)ؚ:5qYTߊxun2K侱OQ)"(Nz N{ jŃew $Bza~wxw5'0hA TB<㲫}Ҝknq+R? %{m#Z9`tLC^ߜ} ]wwahj䥶s ~f7xi;EK!Q쬦.o?#˷<`C31/z4p: gMnNۥLq|/ȣ ] Ǩ eM!B#I^b,Əb5Ly1(.↧L?$C|U>AVWQA`N*+(gmTT@ժYW:]lEÿPϥOKc UKRը}dRQ'Xg"N; ^Hݝc@cc<6BV25/('G´kM2CաFB%ҤR #JbV-Fd-˺;l^x%\4j<]1u&meao>)@h6s&\`"&qhhr1U(F{m!A[vM;HI#!ȃ)Xfv2Zώ64R&Uܽl$gePHʃ~k{oжׯO{gz?xж}mj/sktW֯_]~>4r:1EfMoGŽutaJj+W: &g~?ԃӳɅ<د$ukx˿y VKp6]C7ӰbȘt 吢V 0A Ux&B:͗TArYvZ 0 @J1sbnL̇B˜eFfڽAZKd54̝i`rH&<\!QF1|s棅(9{IGe]̅3:1KIː }P0QqD蒃&\6Nr\mb|O8qq8Ų _sP"PnG!+XPRAaȤ%㰐$1M#-B%gھW1CP_I{YYf@e]:W #wco+j:)KШemgGċAkgfP3]ieNzrL͚t\vMɀWyULҴT4a+v^xiJ#,R 1 qhN̋J 2'إ8%kB)!Yv]RdBRm(O6>Hq08`ֶng7 o-Ϳz>O.<ն pl ,(ǿTI;SM614A*&FFAh}HN+ Ly69 f"`m W(A$F.W*Dyе@KQfŞL0^+*{ibwa732{&6l XoX!A zPI19dw20z[FSTҰ-\tS94Bc dexKv|lYo>'+'ܢpRI[bW'[I5TQdl` 1 20o,[SWO$%B`)2O2`)lX)>Fcalh6´A?L,*ج1ukṈM!"B^#if^,I!u紙.ʂ[=ʚRn?DnMhO#ɹ-b1=^|> ^dN"Lkd>J㻈0RrRZ`KNg T),}dTA&KR7K+64+Zބ*h/Qw*(-TwVYme ;ÊfӲή +Z+nꔖr;5g jl7jd ,`H#2bj͒caΫӞ4-&t OC"Wm'"&I8,"DlLߙE^kw6wy!݌{ٜ=jHG?,nKR.E$`ZF ++ @ص+.DLf֙c1>u öPXUβޫ)[ݐ`7I{Y9dVc=8 lTJH<8Y0!10'QQY vXU [`GUMG\ٻ:WyYC"Yy 0X` VlK8KIQtG$1 I>nXѺ8:d.*Yw "5ϭE6eQ~V1yg8B1u9oP^RgI%ΨyL v RiY~AVj򞻏ݼ"Y".ln0&b5C_T*r|f!/V<+Br#N^b:>bٛ#bEW %$fnmQ0Kx e=lBD1ya IW*I`J'z9Cd[1Pd$Byڢ{]ta8R(i ½/^XFTbrwާ> gP>37d@7Q4+P8CD=1{L$J+198r 1 V[Bb y\capD"^Gb$?`.ߍ$ QQLc-@*EOmFWl6#:gH.C ")ݑ$-qrh#Dye%;-b Mu4 I*j(x":Ey)xCl%is[[38fBU`6J3@^Nw] Z)L, v+Ad6+'A(S 'G$7fgnuQ'qqT2kQER?  OEMQ3]@;vf`_lHmw0GCD;)?wm XXiwWWO Ko vN;%ϓƉK[~zRt?YYK$ȸ8UvM$ωgni{&jR[6|q9s&0}GE ~*q wȜ(a]T,>XFO&sGԒsgGNF{սԬ3{t:)D#hJ-9*hɸCgJc"*4"`k:Z?- R:dKCIX.z9IHôrPT7G_EwJQ򣘂@e2:Qe%[j3{Qmx&Rh˛UhLw?W>tzDهt <%`$KY"g@Vh!.~l`$~5"&;mEp~`RVrOoDA*T5TۀL@qz8W]`{SE:A>[Y/:Ʊ(>paK #FSs:׹ pb I@ G OE7UǝgM<*{Q\ٷ` fnmQ0B5oQE +?ݔ ^8@[5-<ўM8~Ꝏ-7S|mү3?K> > ksq˩؎PaTısh8.8-l^(13XPar_ W*AhzwM,]jB-2*Ɇr,}:@q7j 7=wk:NxET􏸘9ǓFUt#qRGu; fAxlwUg]Gc MκxX^k0GU$gV\ qv.wo!T18ЏБOK=n(Ѥ 2$-YIlC!N,vhLɎ~~ >!ZoAœȇqFŝÎ,9D!9_n#9{cdDSmnw 9ObTnlJhpxa<¹7᱅X(15]`94@+)pxYUdwAr)w7j Hk I4'ـr)m=׆֡X읪pT+m~:z*HFؤوZ]:n⑏@ɍњOj%|M,ݒw+ QI)؉H8.zfWtQ5DP:lxfZB\ q#fOX0:h–$o a6 )T vq}uy{Q.s24H)[>浅2R6o>|7k*ʴϬfR) يgc.=-;gF]}y^QcD]ĮpTc[QWsrH#sYb_ϟ?'͇OόgZsn+Y9p6qZ fv: m*il0TNQ⼶-Ā >з1//>@08ږߝ3y|(kFT[PѓK]`7oÎlY&$j0 1~/[| +=5LrY>4PTQiRE+4ҧTJYXnqLΙ* 1$cj`?eZUEξ?C+M4,P8]^{5^ltֿ{}m&=`i'T?q ~q'(r|T߱љXq/&17<MK`V߫-Aݛ׸[;rS// ̘v? qě "{x&l[{N6tld@MR,"];1emud."ݮ 2:7͗g7-s<Zv{в>Z~/xjgs>~y*Әh渢d뙿|1ܜ{`2ᆯq< Ac ~ +l^yu(J{27kʺeGc_; Gs l" 0̐凐l0 #(V; ~u3S kOg^ȯӘ<`ٔM(}UuD*۸B]O}хp1DŽǗ_`lb7`G ؇p/! r?&EU#0$#ja[(L48.a "]Q:6Niw)!S*jiuOp *4c\ dϝ2T\);AZ69Hlྕ[%28.ȴgM <dŤ3[T1ީ_͸ ֬sA@"|i.oYxö]߰w/tcr&(O +2F2NG,0nֈT\q Q8a޵6n,b<"8Hs6l6ȖF=39ݺXDJM,"@b;Y}ͪb] {,Z]X yb nMT<\?5yf|LBq,˄4jJes8TPvI ,\6q:=3ԪU2  ,eR4'++]jW2B2%rdk);})* )+yʝqٿn94wwy?X\Mﱭs&fU@bsʭ\E>߷Ӫ5]obw߭ͣ>ݳJ.z, TBvȕo&{0{Kr\>hlݺ @ִ[j;ʷh=Ey8ڭ!^[=jiOpܞvkPo`\={Oqhk}8xQx yUS׳%"n`_ԽS6@ϛ|X?*`w=}L2eS2zPnԣObKfmw73z1nwf:]<37yw{-;xjl7 =2zJ93ܙL$d߾ob wcc m*6_/5ۆ@CC$L,I)I*RM4 (Z_yXY ̻6iyƯlOP%I/j!~+Z﷮$_wJC ʘpʅ#Ʉr@s'L`BXKJ"MY* ~)WSkg^xa6\%6bHODisƛcvfk擙)Z̝1|e+6(dN_;ꬍ%Kv_}^??ln9Y^M白:6#nXw,;KEO.Nkj y~r1BY?fq:Xxnkw} [@\Pz~Ǧ\b^Ż<\9x+4_hmckaye5Nr"LNP>@%I# iԁnkfḙ.d8y n͏7Qs7qw))N@YMA77!Es"EBQHZ4lD Iiy %)؅ H Fy'v=K`HA,uK&*(`4n'F3Xhl@K]<[ lf4` zCKse#I4WQ; [j3v `Byq0"d*_dSR̖{ƫi1 (/s%et|+6%0XhXtX\V -'.j9~8d4xx 1g&N{8Lj+?_{!>LIJR"D[=k(I\hV\ m;ޒf̚cb2M\LK\&Es%T9nifۏo~Xjvu^.TSjsHNqDbekL6t=(\{%&;Y e3l@YƲL's#hEo 9Gx(4eP8NG@%"E_kIR 1(M0aY O˵H^Z-*( kTᶳDf E)?hpV28Gq';g,_/5Uq WZ x9\n&ϖ5udOߧmto 00ރ> -Ni䄐m ػQ">ގN)+2kM-S۹fZ ̋/;[Ct^&\Cr%9[CJ 859uݚ4'٧& @+,"r3dc<Q.S ]4pTJ]|}>T.rjr eHAQaVH 1`H EeA0)yTk-,LO6Ёΐ8 0g@a I;d [iBaudanXQ'(s,xd>:鑅O(bع/8zh2'²f`ԓ&;} ^XÅW8uP`a=ِ@2֑A$ztlqD-yՔg5,;xz9h\KF y H#EC"̥n a( R&h(FI$RQUA\ ctMNуibO'=xQsK2@įH[Y㥦 +R%>r]z#Q|g4~%SLcޓ ̄(1aXbQ1'&8'2N|m/n4:W?<;pՃ|arԢݽݻݽ#t0琦Y Q0L+3 T!$s0B@i2p/ߖ߁ɍ[Nr؝Fw7!~nzweo~6v'`fjrbwE揞@Pg*MGω61\iJ@Q*H 1zɾJ∺'GM>L#yb:eo66Xv~daJx<W Ǒ5 Mwzn{@pJa o.$Ɏxԣ(D}Vdnn]ghd,cޗ~sn`XtΩHHI"3ȭ$Q(G6O\ xk,`x%T2(W-qdYMZnr߄$|rl鋚e.ʗEsu8S~l_R<,EY_R a-R<ԨY;T2밷SJ%wA^F'?*;WP*l*DSs wr*YX_")@A/㎁NbAx2 be -ʧB4CyBc#?DZVXCCi5,V!:zyV!z *r{9˟?7*B|lGZ<|Zߢ>}wm pc6[ 7 H T6{p5!jG\G|)Ï!VK أ`^l5+B ~ Ԃ j+kj&-8-6TwƳɨ}GgkIJvۅ}>5ʸD%(g7Z DWV#ӗr!i):;JJD`m4I5IF90] CD38CB_UYr{Lc}/D ^_( .׋r`@6mA`6~ŧ$%Q҉9#R$T x/~c1yUEz ?!ӺLX0`}|a5=}x1wx6Wo8ЦmW:psF k!(P?kE#u` qI: =-cL=+) <VrcLiXvDq|׵o ^TJ BTwૅȕbw8O55 )c QuϦC/ O .%4^uxUQ>URf4G`gJ2Rӌݓ| CNyTM$ֿTHji%̡"S n 2 mSa[F#4fF ΋ͅp~d. & wR_KHz,|b_B ݕ1 =DKOkV^}Oa{^N:"Vrڌ[%ĽKv*8 @{{HeX U&^ cќ!E)l ZHk/Mn GYVW5%(MYz.j{](nkH;N@@ P&y4I,zkc2_=tv4J8[?mSnӢAI5p HQIdϋ=b[#-܂]E *$ y#iʩ(()F.8떪#V C_#|/?}t%d ~W)$7/W0oE]c)GiBmiv0]jP{PcԫoKg;|[sIpADvYA ü97ޤWA !4Tpcm\ 9%XSRyGzd0XFz^]#8RBr ؛ xb4«Ey+_ Qhw,&?}K!y) 1$2*2M! 'ب2]w+wR7k{#'3nj'/uB߿ s$RC) JQ4o fSCU p,7X)Sjtˈ0TɝAF!:{bv-T=٨N(Ȗ(Y+V΂m]ٔ4Kَey #Z'MXI-0y'H>!kWl>[l( Nzd9Ta$}*`jEkmIE0s;3C8$`s @쇥,9k_5)ɴ,ٔ)ʲ,v,W_ $*X #.MV1O\} $L+˓57:nzkʍ9D -f;Q2:5>Ǔ$N,y|EJ}CS]'5?8%z/HZeE)uh)Fq1a)A1k8iXl7P@hz(nP0&L9>׺yUP@(`'(&N|Hu;([vQZD]k5UYm6qҷNbMPq,Dk pw9₢~Zww@iy}%mn 2Kě^d!W&Dznn59WM@zaY\|C6_CPlg*/ӤiyˀkL5K&tGg ]\qT{邼ĕnr]jI.g>އeާr. }H[r{V:߯[5e)lN)btZͶr|6$O*K+/[m@!]ƶ"滒>De7_#^qO_mUגMręö/ri^&$xD R7Zť(oZRd aܑ ,ٶ+(s,eJS*ьI0QG`e[a7CX NtK_ekx i\|ݏ}uG'!lOS]~r߈o ]K̫~8Sͭ2_QgZшd[>*(m<pwx:E), wt5!tW>=Rwk'_ eO8RYb.MTFB! D9o אVd=+z v)թ &, aRG5J[peu@] 7M5$YLjICD+ek;%-zT宧˖Wōcw~y*%9 9{485fɕ6B(I|h:\v#m#bgrz, 8F6ӊZ# dSQ~xtکG|R&8( PBԟc#c+Lnɱ \D#vzi~Vn/p˩9Lfz㇁}xrep/z2p@he٭U@0*3fc9639H0>8֫ޠ7ɞ(8jb""~n{Obx _0f|t4nؖK+Iz(=zJ\u|J*K(0[9o]o?Hs#ڇwo_v^. 9tZ &  Ynx Hw.A]Ylpxt:NNFi%FeEOG_Fw @ixaF?^׿t?0SAurޕ+Ӯ_RuS \W(8v 8_յp]ʮ :pSt3tĝK9PЇ,>x{f|3z?h #E gAA <+s^YfN/>;'N|7ܞzᅵFLz|wܛ\^{vso7El'ղ.(^g Z1ޞN~v=C`䥟p4JiemB^շ0zw \gq$L*̌QMY`KfwE 7pr"綟GǽA@1pΟL.ޛ{8{::QRx%N\0'&.@sNKx_߼ jGz~{{v:x>plBP~p8idS<39>뼩Wc9?_dE)(;(<4Z'P.@O|}>x'o"tvS''Gճ&v.[-~/p/f54ߗ;eejkޙ_Ɓk.DƠrz}Mr!y`MM|ENC C%)&Q9ΐaFql R}%)Fb#YagD(Nh6 e[Ͻ8)@T?e*xd3_{.7@IYahqk&PC2};FB7̺Be](l7֢f)biV"lc! jKӌ8Jq iM+ݰ'Ħ0>eʺ즂Lq,m2 oX~7V1NQw@ ,P5I' Gp˗UyhB Q\ a9( /u߰/cs`{t]p~^Dq~nߋ[8EU-Xd.\n) o1}=k~q/s9k/oc "y&Q.H/S5Sĩ/hQ达3)ʜ$~zmsAO{;&Z|ئDՆڭk1Z0r&֤ f]*7~wTfW_ekRj+EqĚ:Qa1#ޟ qU\Ơt 1hFJ:,`N{ż!*49M"` l_J\PkI4BC@6hǨĩ@C U i6~wJ0C!/N;ax("2 EI I`#`*u (G@'Xj'X`SV>6T>6~g`i%$Jloρ5zWU n:qN77)*$-( E2,(~-{õExr q8nc`tP^U}@Y%/FB9Ac4ki hï}3Il&ٴtͥ~/=5>(w@OӸr<[ ~< Y[q)1wQs=0 1O]Hnn R/(GSj!5Wb ~"T Vwm:VU`M^3Ӥ5q&=Iϻ׉8 Ռ>z Z.p *0)[J#\l[T͇Zv<-HT `7,MIw_2o[stǛރL%TXƮ\[@\ly; ֥G/X m0[3W~j]Aoތ^wWKWDޞpmϙbZ.}h?>k z'W<~+^%%QvLISwRK[WӬ{ RkKo]ŸIsT6Qߵ ,ֆl53.XِZǥWRr0R/2]Uv1k[ W ~كWCR>U{ۘqǯ=dukP'85\JI_ja(UxHHs=rx49Ko]>9޸O*'85Κ_r ,a$qaŽ6^OMN&qA+IےG(E-*@l.e[wV)%]\@rd, : &$Yd.:?{WHԼ#X@? jL]}L dٳ.mbӲ#e1TV3E02"Hةj!WJVlh ƹh ƋD%kM'$H6+JK@"ZCMβTCУёh`8QW*ݒ.YMWn;b"r{HZ̟U#ܖ9>_Q?xgBtǴN})W}0#ٚ7ZF?|(]MgsN4돍ߙX*O_}2ˉa8]MM%1Hqu<&\ 'K154A?H]'Xf}AmV ;$6[X|\o⍟wHHw} Q)&s1 kf3=IIوf%#{2y6$L"PԱЋKL CsԎI N&yx8Wl\,@@AΘhwuRBHXM̂JW9LK:u}|@Ä)VE~>KdW蛫|v^?g>0k j3E>kN:-at:=0yK6ɗeB<7OK9 ?8/񂻮U/X2ț$JmmR]UZ4X$% D p \%,Jߺܵt?˛Kڣ92gY=6мO'5]sanIQ}kMdvk?%¾SځJN}k 5\Z~]w0V,4y< dwsi E! ᚌ@Zp)|64;\9; `*Kn\7K FܧiT.\ʩjB:} Onp+ZP @]D"аb4F JK#4WF g<`ƒ] &G5LpՠiFDp[hLiEd:! lYmiԅNO1 \?nb<9HFGQTję1@+ʅCii7TԂ^ĠcPdʙ Fsb:$rM8FcEi-Q֕ >H qk%ErRrÝ'eo=E-i#HY+i(;ӈz26}t.p>8a 6&N:ى;FM4'9P].5~{/BVYsČǤS4P_ BwJofb"%/mDWYE?ևcԖ<34!<1k.ȡ 5Ӭ]=1h7<&Q-J}x>O~~.öH+nXiI`ϣeqf|-KjOlf<oӠZXNz]u +`y2PWr^Y [- r1Ɠ3FRm)&dUwoNqȭT\95#d'**IH VѭA!ro\_67 k$A\NFJV[޾Z֋,KN MX%& ad{@fYr:,%Ïep`X}QpIy'ba JD3YQkhhGdnȁG0Ԃ*_!`$@F>IUh5F#ȅB9GE/{|F q$`%sQK|nR9P~+ni5D|sty@ERWh\aAA%ٝlJX-=ϝb:lr&hb( 3s E$LEE"HWJȦ.Fa[h?nתm鲂 YƳdŵ-vN;]5!n4VE3$&fҝ6 '680ZvjY4(Baw7J.&j͢'+ې LM+3mm/v֖F $ku@As՚4h2J yФ i+AjwӜ19 95J|pDij [2BrR:5-v9=9=}[\.'ũmtb9Vm DȉRk1h٩-蔖cwF輕b\#LdJ19%DLhD%5T2)o㠊'(@ HIaT|}=].v!WW*c/Λ H,E ƤGb(Nj ޛ t#؍]4m6"*߯W+4Zhq!KnX}ww1]yMDlKpW>.ݎ'HzDWg+LlJն2nlcL_fueY ڒchdqdʹ0v61|rgY;^s-_)yGe?bNcM"6w:VV(fX w$8G(ȨXֻHal?T}]0fv15t@3OtD UH\ KZ۾{(m1'3WjSw+4퐇ЫZA2}nCpܸd,b_~?iW3'gWnX hot. r`]Lg䫊U>Bcrɪv2cV~½i*QmX\q1Xwxv5i1Ys69pw܋x/+>\u=!xTaM?iW]~e6%m 3Ԇ{(."ܥV_U`Df|OG'+TN6ܹ d. 3/|\ ,~>E;CqR1IOtKB0-|0HF)+!1\6DV|IJ}ԖMDeì(bbkVɱr8fUFY~6nQ2֌:)ų^ _sYēǣb 3`魸(8(Yl\C_C {kԪXPZ.ְ—Rʇ@e_N/pk3:⠌p %`,Apxdq5h^UYZˬd/"BnX\٫nv{~~` s"zN4]ΦTwC Ȝx2~~"k܉P674jAѸN:A$󟪛h;c(0ڨt!Yz4W6 P)sXE!NV<# Tq94-ش7lR­x -ÛӬ(Æw<#:"2[j#sgG=xGP](ϧ&bSvǘ,M 9ci^R~QrQKBSZ U0&ً~/Av UV"5QžNaVJvsֆ{Yxz/\t/J}Qȃ=`hO*Р9V_6朖S}NamZ˗L)'U V } -֪+~=v%3J!Z[*xlTL$ g6*|%-ГBZO]z|O?x.7]7o ",ݿm kf˜ﲇk2*Șy3 ~]V?d2=4G;\yl<\ys{g1o^Zi?Y%p`L/'g[.ͻ8si6\o%ˡet_ d á@ v8\ZNNLQ$f+L铏R$7>PKh3AЯde*6z[Z!KI7CdS +ƼC'qIVy٨GS^1\'RjŽ~搥z%v>)X\"ȶt}TЭ[tVЩBt*|?d w[qD W_[x,% <%[ WQO_ F'F#:Y4<}Z &rƪgJ{6_˾]j>vn`#%l6h4q[M(Qj (vSutWWł#^ɀ/Sf]1r&G*w_>=0p|ϻ^є*D^}f<}kJwG c>N_DDh*{}~lXx<z8(M7Ei%?lәy !/" t 6,ndmhxQٸ MҏV>Oc9CsnQgpt&5&B5!mJx2jF7BΓ+|NԻrq" $`+ (f<ׯRKuB.g[Љjk>=H9/q,RLq:*Çŏ^qU (F( MX@b4 1؋1sS>cT{c]}ҋO8/Wf鵟/BH[G9ƌ_/Л]>!:{ ׷ɨ9_9]5A?+Z(\j/vc"(qO#'ZӸSyeEXG0$qbچ  G FEa"md` ac B;9J9+Ȑ.I}V25T_ن?kD CO~1j Lhc^V`]#%x05IPʷk.xiFV:v ",%(vt]ƿ u "@8a`- 8+ EI$*(c_ ":Cmޜ<9FKÙbeFa_~HXSԉOvI^-3$Y'6׉"z|BH?W0—Һ8Vw;Ju0w"Y=W.S9 Q}{KZĿj}r#!R:tzp d29~O+ wx$I}^463i*$ %4i4urid+lb%6B [k=V.L\':eb+6 & NP:idlR-(Xǚ& ձUB3S^&:-g]lꬍQTgI*IǚY (0lDX1Q2I,$IG`nfD#^VeMؖ%/RmHkgzIeDR\-BDGJ .qdMh"VKU(7ɏze0QZHh 0%Ve*"^]Qtw{[Y6^ӗIPQį Y նW]$tj9n2C0/j vr_q7(KP@=FC=xuVќv ruT\4MZW1aƅqOڰŗ n%1L΃'A>C(!򝫭8 ݽhLY)$9S%*,_2ƕCSO&G2:dF {%@+&.Lf{v-pK={u ƔtAh "GjVӨy5p]hˋ߀mz1($jfh/!W%c\9ys•p6 N[ywo'g$i|VAk ʎRp=A'߼,vQ;\R*>=c|`3-$;";)E2GIƚLИ4t#wvXzt>(i6պDwE H0.sg pˋVVJA#18M4}Ү 4j8H.f#wagfw8;4I`x)dC)Cj&n 8Bt7m5-e(r+8xGKp!d9}jF-^6\Ưvְ'N |Z?Waĕ;퍹}?ylO'oq:c]jZ+N fz=/p\"ȌwM'2\tO| u 9-: l8{dʻ煝eZv^U~ٶXaLt1??&grA3;\T"c|%9զ%=')LrviBp0$ T5RPʷA 7&Ҋ'$!!3s#(ÄD#)AcETy䴹Kl#X] Y:E5ASɡniZQX#kb)eY{ݎ /Ȏ ֒CGm #,-!Hi H  T&*FcXXair\:7 LsI1ǧvMu˦Պ+NE;re+8L(bP i#.Ԗ)l,ALʈ!bBl8(g85,^i[*dJY&[j}ͻG&X]{C*?=V5{<`ZdYiඐm,q_̦&,zl LZ| s6.Μ{WSpcO7wZP_ P5؄%wW/N$46cx-}B92|iA+0xͬ)dғ+z.ce.dp;ڍ\[] |DU)ÝinMH тLO"Czpblnxdz$Z Ӯ:3}ϕf50貱kaAo d֬o9@1ƤhcjD\x`aJvK.$w\狋͕M/Fٗ<7@GƆ/c& 術fX&GZd@9ffjKAa)#† !"SPHrsV_3Qɬ8 eIpw olP4>F%ZvP`vp+6ةCKk~Mq@Cd Niou XZ\J°RJ 62YƐJ<(sR3AjMӄہ_xqd"$ƈ+)\7{zȨWx%@bb~m88e'_=xT\fHI=gAqvnrgBnZc^mқcA<: nǣv9?.W..NyN ݺW_oc(tN|TN\vD(5ȇضX$ J' ЖTl/qM> Bp.ތ:@y ¥&mk%[#xH!Ts'4)J)J6Q /M鵲a6M-' W2X(THek^9fU0c{efsp ϋ^7':&7׋sċ1%c\54Pwţ9giz|n}$I8'A( (f)#X9 ]#L|Vyu,Z ]}z{ x+実8)Dl!<Ά7ONbj β?LYg¯~ g佣V!6$gi<8Ǵֈ OŐ`W%Pcbr*s1W. ̓_ÅEx\;4H*-1.? ߫ 0+ z .)"FAZY(N(m$(1FPiC]yIm̔!bm0’[?|Hp[ִpnqt7X6f ; eG[l?#9^֊pu Qw=}Yh wg56Yn#* Ш2++*k5rJAr[(DB#ReX*V(Ji(t*J;)uiJu0;;Jfo{'±aP*PH~y6)TaS{ ]y>nQH7vZc[+ ǦAAyBtQZx6UE1.loiX5Ij ņ1aqDiDLD;MC+<լ>Z;jQr-D` 곂aP%HoV⹳]~^ϟK}. ]G0]G8욀KEcdXZl4b8rǤ+nm՘}Lw?Oa62ǻp$ mr;WZpbmka^DVK&Kr g!U#W5긬A;mdkQ].$R<2.$Ӿ`, TA2NdR]ղRИvܽPEA2WTڇqXISPv5* 54G>9k+eB_ed>b^֕Ci*Y=;ߣKӾ?`O(QTy-˕<_DQ𰝦DV2"2./˩?G@VץA>FSҎw|+OBIs$0c=z3bIf8>mߔj0\妼V" q"]r۱Uu ZgTLQ]r[=R毬IG`:4A4^gKgD| c.K*-v+4^U8sdI Ks'F,(AW< Bnr))`] Q$3x35̌VѠ$2Jph`HFF 9J 4;iq z+ 244{ ?`:W.RS0&W 8aq X%)rv!/͵a HF\N+ooJ2Hs)W& )ŹQ{;9F[Z M`I+Q PSݮB#*xD)JLtU|y6LWsw?ڞH[lû2iMy6+ku| VETWD2ų}䖂":SPeA\2wqyud qacٛјv^K#ꨬ!Xz.DɜEvVxAJҟ*%qʏׯw_'X"P>~߸IԒJ~VnEψc '* +9-4|/ I0JD|ILˎw`E FP Mr4~zjHybNT)ETJ>۲[ ཁf#M;.sn`~[O_]&:7 "Xw?Bѧg)I&1o6܈gyXn^܆x(ǘW|;q9@3>\5>B5?I,Se''.R<{L LIb0QnoӐ(7zpn&Dt24GgK־ZQ<|>u9ds5pt uЃWa J;Pp)K_#??U몕a-i4E4; >QLY9} r x_*HVfYx-p$ sLr9W.4ZuN+JI35Y Vw?t^Ġl5ڇ].9q%&Z-d]9#k ߍG՟?AL%ѯ[D魙? %W`y?6 牄 `wR.W=:!RhXYy!֓+>[ 3`Q]J]tq!vln0"/J+8B Ĵ`a8l+6|??;D)A1`f ëebs *yb0>Ww_sNs_O 9뤮e=,\a dv.iQ$/z;ޱc?8cOsOgIꌽD5|Ȝe_.{JBXk|dBZXRZ~?/>8B)rTI,;zy[v>٫sA=:BgX)K"a[Vz#36`k]k'au@$i3oŚ\=y\CiCwkIу [F|pz޿>7}plXiNtz*න P0bq_ JiEB O; ?!%Uwg= :~G.]nRvDn?V.,_-,PQ,, sgTV,pq1~|pȋ<>͇?fQN߲1|?/ qpGLOgGkLqjghΫ>_Ь+b|7z=*&cD9"HC*S6Ǻtl[7Qq@떋A侣u;;@z0huBC*S"QDHd_UAmĺ"XQXR&,8 ]]^"O,1v_2bcل![`|5OslHjƀM_è@ ™V[xd,Q-)ad-Ão-ӛK5X ku!.{&Ff{Z|os*F#p)?vӿdTcx{=ǵf@RYm;dBTQFlOo~7|]Vea/+67fGy9O+){*L4r8ۑ=Ws*e"n LJOGufb|<\۳ū@RʳŻ9K7̧]|W̫ 2qcӮ~l*VQr aqnUx'>וV^0Hn(Zoׇ\!Lnfs1GOſcÇFh=;Ѹ.Oq- O C|?=A Ruo*t-&9Cf+1wRX0ĪY@k h܆dqa:"bL0#K}AEw7v nnpQ5yv{. .ۼӉ"D't]q;V6p]םΛNdm )O)b1i.$ =x_ Hcj<ത'ØA.5(+ p¤`^:f C#% JE $% K4R $(KoU$QF`&i%O5Ӑ\kM;ESBp00)ʊhINma(b5- 1D"A7O4?$ O$4[-(p8!HOI [+v- ݦ1$@F֚:.>7f<᧸v_ߕ< 7},ཏQPŗW~2(v`k==UK)e^;+"RSn'Z1 Kiv۶Pk-چcdA\C͕zjjp>Ĕ|ؤt,緽<* q*cYԺm| &igg5֎>~y3*!*ڟ?YLc"1Yk =UyfMDtSҬm1.~MmT [`q/D޵q#"e]k4i <-^m%hdm85WtǙXENDɖw,,و(yj *H\q@( rq37qc6\ӴC@ZGr.$BO/v/9)$I+%9xvTr=E7тى N"*'[ޱԣlђm7+Uy$E\mp4)P~>L V9,c\qbUPDC)&/Z^{k%ohPTH[Gzl׹{ /bi & 9Vy`^/k;"Qê/wS_= 5_-{IK[U4=[QrÆp?J\HVFjqp^͋9в<8oô$|v- ׃#?E ~88{uF*2A:0gO5/@flͺPϏ9F'F)cՈv+;X|!b[#)L8Mܧ-^d\lj}JhhWEo[EciU<2h}Fcƕ4O[1f6o=W[5¼ޠ-9I$ʡ7?jU宸D+27T!+Qmlg, "~j?zWunܖ-%pU$[Y[rpEFu+eJ,alerӃY]>3mm.?勏0(/vpsw+W+ Nw1p/m?ݢmxtp604\(Dv*C- o8TUp$^R~Xt(=UAQ3(lTau &&fo`nRh$G#C̸ϺR0Sd|s^A5p}C ސI"+&a'2UN=ySO^ԓWԓO!j+ccW Iw%ypη%02d¤¤51YK +Z%eD"٦pF F6J{B_*C ->hIZǁS2gEVKAjlj;( B̼ Vv؆L)ͨۊ'@pyr˵3F6.C$y9D2h%'z31$ЂiyB(G;p1\FƒѸP !Ĺsj.#(0_N~qݏ?@Ɣ h z"D9H>1K_(S@TE@ X@'+! M_w )ú 5W* qd*K|R3yaV:|Js@-dDQ4E0Ʊ7Xq%Lȣh^Y_ͥCZ2A[2I&at"pKh>/@z6c֘$,:`ERTxmrJKc Xܡ[A 5K=R8fm/E.8ًtV|<3KPv=maEw]8[QGM2JMݱ<@^1ڬU)5ʸ"θlgg\8wE?{q;8# NG*(<ɊQIW2؍^0jgj|Xx1fK[ n[v#0.$%8Q ފ_jL@>;ån8ܶ-N0zp7Sb\ێ9Z"kw(5y#7ū* ٭j*fY訢5 _p M=8fTt30u8Tn]u7"8J=zzZlΟ|Q,R ͥއvS aDăg?R"Ku3^E/ cj &8vwZ׈%1M;ʳz@`у+-|Vi{ReTl)+aD`{J& p%s/ OQ,!NHcVJ]0Bڠ k^,"yͰ7Gv{-Gy3 ef ;\`_3p:3? 7R_>_|q{{E.TY61|}s^/^:_E>0Pkwv *p{שs؅+JNBD/.\lgӚoT.$@ŏF` 8UMuX޲F}Le9m40\ߣppmob_n}׵|uSqaK%x( %9N6=KR6Ey>T3K95x ")/&y?jJZ4Sl T^x 4O%KZe2K"CѬTDH'8XƽAGI ,(<\RSጠUs W3)2'&((xKꕵĐ@Ĕd'h"AJVz֥dIЗfZwF#l6oIv̐S &X8a!F騙J2#<Ȉ#VQ9m7wO0 .'#Rz(RZy;MsD#S˧ʦ@+ߛ\Ev^l )$&z&d9է6l@mב$bT[({'3 <%3!(;Cd558ְ b1IhaV |{0AyhR9)3t$І8?BN6yZ%֎FP^wSLs6׃@ȢAhC .qechcy0)j[0+5JdDl\ ^,j[O.Jhg=~ i)h0?UzcZ 'br-/hH֯dŁ))e"0}ap{^yóSI>*ʶa(Et@8T S"E xԉc YNMnP]YKp95Kvp٪b"8WuQ1=!ܠj(P(; S)_ \NTlHj&V7Z٩ V(N SI)+ʼn"x=MHCGɝlP/D\rG:A@ˍ\Zj&YGU<wDm}rDw6qtd#[i"ǘIc( 0O_bg,:ȩI萵m/W8a d*Kw@MJ#X{ץs8?9D dg/B:ݾp7ӟ's"K2Jf?Uy|o?=FT7h 翼ֲzUy].p$'FaWvÀeJpp?D[T #:]Ht\ݦXݑccAOrZwjqeN4d87Xꁈ[\MѸ{`!`=Ҳblu$AvǐZ~=b9բyIsFfF};x sw5둍#J5nչUA9\L׷rVԒbܳ,ۧKyoL(yGTfE"8хwAnMcQ f܂WmS_g-`;g(\mpTI^}-otF"r hu\.& T#V"k52t1/ CF~it4#d|f#g|?\vYW{;ӭTkk,Dc\磟V*瀓#`qYfbQMrzsQoa* DP)A =^>\Q'Oڕ=|.bfIQ zݩӢfgZS3^7TX}i"SDk&КggFZ3-EbEٜ=rVT#9$|dC_#0ߡcf$wa?{6/[Suh~IUv[[gOه3[*cml9#N濟%;LJ@RL&ch|hɉoRi.؉O}S{{>svc4!#|rMB)=bTrXc|s2L qΥIRXг3 E(a [(Tb ЬB{' ;T/ߓg~~Nz-AV: !EKsT3@Qy,*$vd!'A h?Yq,Cd1{tRt/Ks+Cc#-ɥAR1s(Jc.Z)e 2hJ)Iڛ < :MA[:@a\_{itNJ Ʊ'@9s 2V!IsdH25Q+qړDiC`a+P6҄:*YaǚpR['e6 z`Z f\^}+o ~엁/%U ꫿T-Ro:T~7gj Ld`q}_[:avE\Xǚ*Kkqp*YC_5/ga6):Z4/+ d{̯-ag GM9pnuLafBc&=.qq\8W777k7Jjus7|f&+qD)} ,Prq\rgEfަ,uynZ.hWOT`dX2U5"i0*pYk$q-IM7o/\ yn6aߖ]x:YEv1}JX&'[WOov)ؼz>  *>d!PsMVgpC/bBna4f8: OY5@*%"Ht.}ʆյ!i-bq#QgY1O g,"ByDzj@ Vkot%]+>Eŗ1~TEqeiш7s@amm/vYf蛠4S7&ﭲ7Z/<eIhK22g %(aw"ǣvK+I*r9 {}lju.$-p+.!1VuZ30+J´OP!z~Q844|=e\*a a3tM6A]6դMx%9Kw*ℨј=| Ҟ ~K䑮6|N3^ܿN{=/vxB1&}>jCөLpJW(U+~ REr945,M ZM,߳+4٠3B޾ OGzw[Bo8SiDW >ߖ+Xe~>1$A7Qug2>h3_I~Gi]>M"^Nmu`ЍM~|~pe\39:'>0L ,5EQeʜajy\*W"Jn0 KUÙ1Β=a6ⷵ{0h';7ڙkk.!.(J:mD>?Qa2TD!9E}xU!?~݂J8R=t{Nh bTzv6yLfaCY &ao)kQF` # ս+c)1#2DIyC{VȿEqaddfeەRyyrA񅿾Ϟ80i+.^'ߒo%LCsE?ЧSY/G+Bp&QR?aNPW/^N!'`a3C3=XPBqc . ?,_(0N^[^_ku:?Y~R+2" IU,T36SY efT v wϤJAEjئ(M]1 8e6^EY c^sAT uՙv"Hb@Bor|$iE4FrE$t38e ߂[K吾J;.j ]L7[;\mP5D3HՋ9.\ITЊJ)XOlrZLC9 ᘜ<Cƨrc9S٤%DXC_y-oФT/e0oA.W [5$*LCi3 vJ;7w c,^8pr  uAl7‹P#d~OMSuqBJH c+F)e -Y@Vnq;ӆ1Y V\cb)H\U K[VƴD2"AM&)(y-@ 52ں6$aXm"+hF&xWb`g(>Tvm"Eǚ 0`z&2^ Ϛz0P<-t1}.Oi>^L’xQ8b1E^Jf4l(N+y a¤liDP$N(Nx Jq<4FC)/; z)n8 qiNCpZn-;D#`J+ɕ`,*؝%go`#C@,тJ]@i.qE,=|y5{z~NCDu"QV#U&jP:0QVz6c%`Q#E;ňKE$/$#K(#Iw=5kRo`iDFsD jb9Bs+c4ʥFH,v/69u%`rs z+1Vhƥ+$-%F3k6.,3v&4[>,ZfZr,Bq =~>[.Ke,2/~&WAMrnoiZK& ҁ^?zr"LqnMIl ȸ[ϫEQ"*RBQaPpr(9 )ęcÝZ4;mchMmN SSpncJ0XI]=*/B4(eT'#JU8g^^ %ik>?"G K譵($a0]\SVp|Rp \ ۢ$ne(ZE\&wZ-?ڂo=ީЎzs}ʊM8d rVR'#-S n‚/Xe<9yA r!S*C<7fAǷ@>(5mRϽ&Դ$ P[RK wx 0ʌ9A2VZ Wh4!TP fFKg=r(ʵ8ee A25֘dJ!68+Dn{z趐ը! I{ cKThtXWךpalx=rC%h\"j'rVcb 8l90%mЖ*tu .4bk/4}~ߕƨvWﻞ4?14!¡G zQ" qݚW/Ckz| '?=?n&yl2J,b4YO>>)GPm/7d9Y?ɐt N OFlsɉl+(:ߔvafM(%?Ďtiu"9QmIRv#9ZqumI<FF$HۖcH#ՔԸ n N,[ahu1ty׉; 4,V0y)O >Fjp >W ɳ۞qZsU %Xդ*9п,˃텐YLBXg 1 N" '׈L `YU*翬<5.ZW!Z OP=+c 3Y{Ig^ }1*y3+(BIUE,o<{xss#U&ԅ2(GBA vsи{%`Hkg0 71&'WPN-FG#1#ж"^>?TN+DEmf8>0Ej̋lȖ'^d˓,28#cFm~<86㞀I+][o8+AbU!yX<4j tcS3/3tw;NV.K9NIJMY.fU"}yxndD$ײAy;1\H\At6M~Z _ij.+CR1NɞlhDȰ$RH)aQtR|=L=k֒xƲaxjGjEE$'`H`lU(9fr!I X%I" jcBuf-\_lkqQt0Ɏ {Ç&1ABJ#yopz&^W9):_ Wy19 0:05m@G+ǠsB)cJ_?V|(Yl1r;*A1Х']nJ n^?h.Q=6mוFxZ%(kITORCQEN](?~>"JQðgYю[х5WF{$D&ox%n~ƥY7C壧OHGuX|+Z*;&ljM̺@[3 gf?vP3akP _54*m`*j6_A{l0u%EGz&tz/; VCVxQO&Ū/~IpzK>FhhhwBl!u7S>7&%2}. 2{qO~Z-/mRgI߼{^v9Ջ@orޮMRt%(>>^F鋮el#MDo[*i~%Em"U/~8\@)QZiQ[Wvn!Vuk š9Q.5`N@ԗR}uax=K*ċǎ xwL8vlh1PJKE1֣ˈ ]G9 AEw.[{9$QNA}UN8Cb:ݪs1J%vﵒ xBFBUQY2dRemܷD+4<0ᣒ fpl!hwֽ"kl_x ' 1ښFSэD׷"k7>^?||Ttjuo'l :\|UdaUYe4P|,d^ݨZ$PX&W/2+ULԣEqmr"&BSniUr.fƾ7TUaizJt,pԝn*C0ylܳ 1싕0#Z3q$\tQh5Fgȭt NLTdL>:yQT# MܘE,_65cc)>ZjE$<,jEBJl{uJ^`Ji~FI`,BQW&P.q8 r.GKRge40Qb Wʥ`{U+j)PȾ $H0D`5jxDrI$ٺ[`Z=KeN_UӮnNVWdц JqG3P8InS_{k=^^׾`˱KbONoŞFPgTD+Fꫜ  Oh`9Zld󴸓x=kbi#nCv?i$O;iFh[3XHL%B`W)U"ID.E-˿=68&"Ȥ~ԌuJ,ɎJ A.UN35NyHՑba{y@9nT'HS(?(W PC;R*g^RT|!Anaf1;kG?[4aB[{g&\=gi"3vMjBT3=e) EG׫/-ny5]–NjDs2 (0p{UeJt,FlyVK!S(2Ԫ$FAEܠRDDZGkO:a{+X)Y5Z^5ǎw l)CJ[RfUʢXXrUAPyp;f^*;,\AWlMGilpFvPf@Σ`V#f$Oc?P_Sa Lٴ7;@g@];4Nt c9]^C/[wi+T; ϋ W;x5烳t;FhEsEpwL5+s<5ĉMĊ"p"4kآoܫj7wDŽw@7"% zךH9,\ ХۜwA0\A#8JyE);&έWqy* B>MϦ o摜Ґ 1zPPv5ir=QcΡzևZY*Ӑ 0IQztr:z' jH؂ЎWY^D+l ^a;B0C)k& cZmR=n}Gvj̹dQUZl}tԙyBڋ;;&+!C2m#jn&ТQ!gi5X -'e?@oEH+L(udU ;4bS =Jb|zLі lO[Fwk *櫨LMrmOG birfTI8D~\&{ʺqyM 1NlU =#N+1CNMreeFR^(FDWCᨶjjF̉=6&.'A;q TN z'ѓ>``=&) (P>+g-tkD:F{PV|7A.Sӂ`m&yRv~)yaJA.UCh $^s?KnGuO)?..ᄏY*# Mwh$:;cޯĐhk?3 < [3 䧘Ocd*-l#&i70hlI@"SF1͒TcMƗ*\^䲺|Zs-spfL†Z򨈐*>X&tbf;N*y #UOr79TS`H,)ixS Po9C1q\hNXHg,qyuanمYzVyQ\K̊Y/_Θ=&JnoUaY%\VK/a-rP@e΅K3}K f2%TM3 YI}=]Ռ$qN(K3c4TnVKˏ^Šb8U|3FQ;~F`7WB^8 1t2 jVa>Ea0Y/` ፔ?Yv$hf>oGVu ڇ?χ7<829ΨqF|S&OJ?Pֺ6+_ǣV<佛)^4ʽsg}LbFe3*QČIL56b"皉q^ue@ כx8^TMHS-"yoީ-%LHc\4%ӠCt?Be P <ŝ}T%>ti A6Ҿi S \ n&(JPB;<$M T`\Xl*k`V㥔"KJc#hT[O T . $52` ՘K QO5xYAp"AAI%-eBdUct"JmXBܦNbv`|sHP0aFJ1;;amYA! #= ".Pa^3$'Ѵ rlގ <M] )6ka&H=Ux4 qMT5} >SHNr>)9@!:IYٔ2jۻKQň|Ie*kn0!\A g㫏XF0?%oNe8OkO9#E PϟN n#&\ Zfh7!Ti^QSvΏΡ-2k cя+|Axo=/|{Tj 5X,fP,/5VRJO c9 f`հ8{S ӥPuW!t::: Ϗ' *~=}.H1X0Rq3dȩr{*\_| ?JPJ-1>I:.' 1^ݳž"g 2@},2<cܥ;njY W&\+[!|XnkBxuk\n]azbEX)6wx†$_ykv*.|t5 d)gЊZL |+*}f=b/̆+!zJLRnH6[)fޗ[T]/aXWވQ,> u7z5Kr&n:W{L6SC޸t_@hBŒ0 s=בܾ^JFb{7Cz-/6(rx(v:2peTKݹ{7#=}mY6U DhmN3q}#HqYjȉXnǷvaϣm @S~=77=5(@|?x;|4Bள!NA]=M"2jU=9g*{6o53Kpq<-P彿wssy2ˇm!N{vN=#ӹPy;7<}!F_?;F!#Wl6),':n9׌mIӮ;.)dU /SAK餵!ɚ?k}ʖFS4kGjt:) RWMkSnDRbF^(BHJګ5v4yqôMb}$X #HJp*‰S&X] l)MS%=5LmcQ3 ;] G4"$3sMHvHExz/(fyإ{<ڋ6^Lհ H-Uy!7ݖ^ ng qݹW]TSD^&ǵւ @H0 +8ON#Pt̮8%:lŧ9go*Q,!;UWfĘ |mL"A6m$1:mwQ{Ipq/M9籆>LQcTWQ0@Z} , ,8A,h! 4d}0a ^ 0N ` `ݱЀrj 1Jc˨F8AA<8Sr2."'-obq0M#=щ5fP~.^56;;5V3==0jBقѳ?h򱓲 ,!S.g>`sPZ)MMp`g-%dcR(A> ;@6fMצ7+p"} \ұBs蘬:wiu|nm&aq9|Վ(*4D;tu֥2&<*V 2pgB,-y2eDI(@iкEuұB Z EsX/V=ʐT!\$CdqM9 ɝP<'Cx, ف\{jlH97%,vYD~&e3JϜQ:#="E4g ϗ I* 8x,Jy"^Uz&8ιH2F#@ Mܖ顱7v {hxSB`Fˀ@nN!Uz<7 ve%o1^R>.h) ODTddeGQɲհzJnIdBҭ-A,ż"!* ZCeKWV p`etiA KUHRhCL;x..F7\y"7~.TuÍr A!Fx6*lqA!╩"TYi0,55 A6iYy2t:]Nz]ϵ`LL*G{&ZpC2# ZH 3 kVooj,f# ^/T(qvsYXrgܸ~tᵱ xIumɛ?@z7@3eԙC?ӭΚ !fc5~&3@V oZ xաO:!I; u @^X1Ld$%HQ%@NԀ_ir;/5fXYC."WSչlNbwk5_c-w9KRTNbUV8|g۹|Hf oe1D5-(g~u8tA _&kJ cIni|{;-/`ZeO:@y6ۇ5wz?[jIG\K~ {uƎ{_`ћfс@7ھom#{?{{ǟ3kpR7R ܥ u#pH4HB lp 8C_ŇgJ  א@ffk?svz㻷?"|zi+}8dk~=~xkO>nŋ~?t1:^?}1ߏ=>8>ϧw~AfO>kXyO;kxS2?iua$ˇrϨ-e:oz~?ѵqf_vnU^ro< q!a:W'4e`N lN2]WTA?c轏99t '߽n' baDoFti5]]xR}T50.MvbnLl5M?};vA쿁/|6o]y7|=o7h>~zit2n~{z}?V#N{ִx7{?{۶/vt~)mc, E&i[Jrv3.DJCrfH*¶(<Խ@3sr:$/5a+t;?LdՅ_ux۷Zn>()_&=jTpV궯RݻyrYXߡ]]7duTx#5OQ?nVӿT|ULoz7#nʋxu ߡǂ +%J_B΋h>6-_Ƴm{r<Ԝ  V5A-?G#s[vlN:˴-GWFzʁ}'SՇ%jA!Q懜sk[0aW /&_L3A酆_!͸3 1{1DL%0ryvRpf0gi6%TIJ [ɜag,2lt3 9=ޥ;D9+tqpbL`Y& Ba4~TEؙjxYԔgŚ `.zd݇_~OO\' ["-J-uW{'ڌctnP(+)[?z[]jڿ;cfѠۺX٢촀_*vpiB2RZIZב%Uob]UCN(@:gD:)8Buk}=#fgRAK.H~$U|xFKGgU1NR"$Dn`NA'aq„<ġd) ιl ӝݛ1]i|5Yl;~f@ֵn2a/'O|ho"ˆߩ͏D)R9 6f>.(YGl7\4+U@+,9k v]8vt0 !]Tغ>n`K *ulwaN1!Ɵyr):xkCɨS˧$l梍oD$eݛ A( =ռMI!DrvVqè:h[oWf۟g*L\wR%g槲FT@L:4c.|dp%b0qzv 0@Π@$U}^HN=I^lzSq~f :ϾsԯZ3Z i?u#W͗SZ,O5^e#09NC/3arv ts[@֥/ g-ܒ ]C]*3ByG/ X5␔❕j>s)݉Nj5D7jmN/$n wu^<.U~߇~Q8PZRc`6Ȇ),x&]1CMrbDɗ?e8U&U i򁲲Z~mQ+o=EVm뛐.ORbBͼ0_gË^./~P[t"ܻPӿ^|/II1 AC$E) R!1zTksa\Jh~g 8wK m w)BLlzl#S*<J SRFqPϤE1Lw@Db Rdt6Rvҹuv} =JF / qD`PqaHɷJQfj`5@N6uMB#x$"!6ĕ $mVC M17`p3 9-e\8jjkhmjӈ'5NjI?PclhJXw)&S~CfM—tJ{o! @>)j*}h: _:-7U\Ꝿ ߙSH9sq8oqw.Ͱ%ۀQ.EaBWun[u@0d]!8y @cihOÉq6gwzY]8nqC6x>^ ܽX=}1A170^_j-DdL`G_wWk_blC4R1Fؼ+I.V?J~GuAXtyXJz%˒;WJD}7JQuG]buAZ).wZZ쳩+,w{ ;(x$9S2r%NE@@d1p$x&b/[aO/bR-vmw&}oC@P J"v6fi6&w̪Fak0$c,s5nl</"+WW E( Kds#b c!Ic)w3Mˏ]ҹQ3KwΓwgo14mu3 9_đhVCލp +XB7r?|3e-כ?ap imh D]A$em+dnI_H, =tC@=΂y\^^tOv^t@; իrUŦ V3*GF[^-^3jmfbSF 3zaTrG.6|]^!= v;pO\qwc1ì\᝹ꚭSkY)>fCSn]~OSY'ef,2Z󇒈G\0 C@Fĩz< #FR DH= tTq+] -?4&%, !C$b.R'2aJ $X oP,ÂF1 Y!Y :(?;6]ڏTzp~8r5UV'G+ۇW{wN!!c}1'j3zaTbi mZ`@=%RwԟAd(BT^D(Tp~r-y{y߃l`hԲ;]Y{ T65hپX:4# c+o =E٧P]'B9mt'>|>3_u߂MuGK!4 =Mw?VGyr>N.a@}xpEA^.Pѻߏ^}͉,ocRgG(Wź=\'Gy_t&ѝt8 Γpu륛mU/=PoZ-'S=܍t1ɑ{&e8iSi+S: T0,>z l Nz8W'v\b"N#*b?A9i9͇Շwzj&lD*lm[1fLrbMbAɩ~;]a}skһ˝Khߛ7Wq/8W_qmj/SS ò?X Ͻj9dcO{vrQc'ڱW~s`dY5=~|'j`SY2[Lo-[E n'WʅW]l;֏Ct~P?{.Gb1T0HK=H: *?]Zvز<2oFƅfxñ^OTG|0F:7ؚ{bZĤx'>׶_ u1[MZ3[I{VB.gY=w.'n΋ް6|Jqv<=M>w@UW~'[&>[/\t1tI|<އH 6OHImއV`@X|bl44Q^HO'2t>~u{o{Ku:!DŽ܋ʁMQ8:VHΣcm<ֈ?vsO/#jWKT9 8{8`0c$IÀcL0p@&/EђGWr5<0tl>jKe4U-_]寛Fh֗WUZ4^ -7,XM`piyJRorPitwޕw(9SɗK3出΍@>>E%?rg?:逋b-̊b-z.Wy:.S:9vH\`[! xFuB)=Uö_NR/96;rFjF9`yy 7ݬzqqB͇1?x~St")RT7f/&[f<\Lp>B,[/v-KVC*(Ev`\&0kmnVٛ QR\TRi5E2$G\ۘ!)JCPWL\69xnXZ .ޚͻ?:ԝ;\%0Ó=_"[ l{8ew8eBlUH-W Rwj(垰_9\ RI )Ƈj =_T IհlM-_1;롆k1R=юGJ9z.1vG;C.ɋU%S2 H|Tq7sU9l ח!RŨq,JL(NIhcke ۋ'%1ŒWe<*\}w\o ag0NU$Ʃx gךˆ~s"]X _H:87I j;-f\:?dzr. h!(JHSc)Wiʙ3~>~ A0 I|Zm[zi3D*8'/Q*Lrl+ oT~e8x5NuKO~LdY?dEF?CW^| *TҥS?=_ g_^O^4I(?Ϛ}'liӧS (N52ռW~ p9~=ϡϋ tEEff;#xJ,o߶e"lg0wV& NH!-10d/!ZSF.`ᨵ"l=:z瘾ԒCZ[<,Xh).#!bP+N$ CB6Z)Vn# 4DP"t 3`8hEpiG/BBcN@&^qhm&ҟ#l""RXF6(1mQ=`dp'-L ȃre lX\Sj|WE`>C]-p PQO| DĒINFYɉH4)1*E \JxVz˹Nes9 ǣs"\f:wt;4W_Yt&n9}x妣Vw]b)+UF #w>;/'7G:X@?"_W/fŌq3Ϊ8f ` XrTzL)\S8LŤr,TZ eH~K)-+]dF)gNզ:@ҦpL bM2;P)ko\1%U,M Ï 8g)Yc|Vc\m"0z8¢G@Q` 6-8kvpӂ5ᦅ߻ 3 MőX.TӪx{<[\R/ݟIs5˘,&Эsa=Lu ~L B0g`W^ 6`E@ԃ-g8 FTk:.  FD:0T 9#RpPr8g5AN_ \ҍv(}>KnUf;nǘ"q@" 9a[M1)2y>ooɻw ""C4UU_<(5LgjxY;+Ί?zş;}3yw3^]> "L TBzsqOLi~g +΁)S)W0&YZk`ܗY\/u:DC3et:D20΢IsS!lQ'3h#Ї":,L+"FD! !|olQ!m,rBFC/EY/- ɒ끛4b9*TwI,GQ4;/]K\t:PLsW&Ryܝg[r̆ƸzYT똢(+0 w57!@MF@"+"3 2#DkoƤ{O4۠ *QGYSƼ##i2Wtg*dd `X>l1)`>-rH)VZ2 e9 ?*ջOP=~(DjCaz2# gzTVa]u"4ViS$QkX B5 Fd|!yk6`B3ŠdWZnޔ"00%Nt{}uz_N_mskW%?ʭ,;]GNYe̪Y>b,`cA gOAKq2ʵmÄ#`"\,h0IU)rR,r ,&牷s msV ^$*K m#xu~hܭ K8_E !# 9b$7)S7,L< i zˊ1z.Eq*1c.3\,_%Z:d- ZL xḶ 3Ga`-NG< a7AKK![nw^-v в++ӣmҚnqT5渻 ykg]mõ%LWaZ*70];S7s-XDz0+^'?v i Va[[ XR{)|z"a`i$-=+"ŖXskfO>*ht.FF‚Yl+wޭ7p҆;[9HgifJ7T2lՋ_kpQ:oe^k6lptլ)cw$#\KN#/悭Z7徳z=4r{8#JY{wu"D:9g7(\kT0HW^{ %pQϰ`)G*"2Xf궃 cB(i8[+*K.Zב*F{ək0fNC evL: B v}ʧχ>HD.yȯ9֛ؠp%ٚst;Aqf8c2J .bͱ.&:=k_Ձd6:TGAYyHFkxwW9 1ء#^zʈVneXhg;TvJ8D f9c#s"9&j4Th 6x9nP\~>\bqA6WbL"\Y?4 [ݭ 5mT`m]HWP'q?%ZTKlkJ ݓ- "X L{6P`e [¸&s ̛He;CRQY8U'M{ +DjjzH M7l&HmT!S(w4n;~ʩ<x:$K3c2WvޮǺ`,~<%JB:$i_4I"95B5?[@FPFnVS65?p\TPA$5h8r]\H0sr]cj)FP:Ҥ7n: oȔSSPCk6z ?#քcj{H_A崻StXN#ouҜR2]#P{-t6SQae5R:v2]uЃt^#TB\]v᎐K# 7^ żAZͬ'kOӑ{$2-HD0\rC<䩔 FA{M(6YnLE 9c`Qසe\*Փl?Ix6wL`ܝcPplTs,-XqS \ H`ܼ՚_o~{J4(cC FN+Pi2 NFӜ *yt2-I$`P<\%⨯X&pl9+7`텲z=$Sm؞O"s4#bI59bFHr@, Z&Z8bL$)Q4{! !@wTQizЭ/r4m%<,C|ӸqB+ JC8H x<Xh^IiXTY A8Q}@܊Lb!v+ jVcZlB sMӔ$*z{z֬t_riܦm]ޖs>FtƂ+4NqmbLe4KS) HXz꽠|zjF'{18Ӛ` gʨ ƌ$Q :B7xI(1IL \Y'01nf!C2(`[D`91 {Sxzx:hi&1a+K#!0%rZ{/-0zuT2x07R-aARn%hy="ԫ '3_mU^dKD`pF@qA`VQau)@F |yg:ۘ0PBB `*X#cXz"1Pձ]|y1{ ﷇNUxXX'B~$= .Zzw/vTQZs/(VSNHӁ"x{n9IĐb{rM }vWUvJTnV8VT+vX GebBƭdjٴ~:Zy16ZoT6<ENenӴ~;/dgdpZGMb֎²TȫȻT>0$NaUk8&ΔW Xa>7]ucW&s.Tv3p;/e1#-Y9>{ɥR*yhQyUm%GߛTF'8AC)ׅk~bc%ǟ>kn.UBtVoۥ u Ċ(R—Ma1:UqsBeCԨ.1 =G1eǒXobHy1Ѥ2̩ C6mC@Egi:lO>j -"-!M5o]V:0w7[~ XSvZUvUMmZj+_ (%2Y$% "  k 18gA1˩㊚s5a-[ |J|~~oNvPkrT{۟K 7F#ouݬƔ=%XF96*5DiSx@2TtOeE3Vg!5y>(`\<編)~M0gH cx\S$%~tz7J-H?|Uf{ qX\rì'D)ҍ-gU*`*)HS/"ܿ~u'ئD*+W.g565u!J򪣉Tޅq:]ag3Zlt[rխl5Q)ۣK̟3) cX֙ʂ>~O+~l>MXsIqa3s%jEƙڋ\!}.'ZZQL9,TV].4^>:K;fO7ѲPY'um{[yY VS= ż)d_#՟NNZC] /Q3}N'of\˝/M3!LjX6roHHo3F5M<ަ|5P;R[*WLv6޲R]Fo)JQo:B%66ʿJq@i7RP +9 J\.l8= z"n26Sێ^25+eeAhTn-x=e=*knT5.a:K q_5VDj0ә>o>Ί+&<t5EKq@qş0[2=$(!~V b^Y78LPH1P)M]@AvH5*gdD8"U)UZ#ukIv+|ڲC4'>q9dJ މW*;"2 ࣹ'Zu%'`R5=?.X{|AU+gbO4e<\lf&֊p!@.E$* N.F >"D[RVh&T i+DN (U2{QCƉّ-:n(uR-09"6V qLK K#gY%;jFYzg/1㸟3dJr̾TI=JYwyH@ˑtwGӁi<]lަ#v61b+5t|Zu8Wb=_es >rH>޼|Qb,ȂlDp " D{wT|ts0< ׿<+"_ tKf}[HKt~DH%pX),t\q,@j"5[t9JLUEEJ z.'. | gлo*‘ߌ&g =8&hm{_XqSSjOS;Ikon>5RB/@7}Ĕ1lm1 ϼd17# U==F߭{[gڪ2 `hķq]ZuϿ~NwW| զZS(5RHi Q U}_K0I͚D֣Q)'S*_?ݬ@k.S^@er ?ie <=`yׯxy~q'GS_꺄n q]+FV&j4ȗ]1̧RF⊎vt!aiE.{U4C3J(ƲgzG_=}i2Yv9y8YdwsYFgf俟nIв(7ŋHfRUUciL3 WޫcP<2 b~+dwT*"EA}Xb~@2Rj(v& @<=bEw(L yQe;S)φ&IUfb[RI}"?å6_RgzH)m1zPWn%T!ה؉Bc`ޝMjC%ELQ-Ҿȟ *ݴh<#a/4S9:HşoۇQLprId LjW$B"=w;^-%iZDl$I(">9O-R44BXfWHKe@YR#(Ssk&-8U}+ SN0 Wk!'[2YN7ᡁ˹$s=؋Oӌ 3*ث0c3v݂B=*4 Uh 5Z1洢iAp,(drmR9q e9^ߒ|Bvd'~Tx/~F;q˧ŵ?|V)%lwn!!~۝F5m9\*  X3۹m Ri-*)p]m~S\wCJU)g=j7ն,2IoW+ca*,o\fW$Ua.]L7G +TT7gjÞ]6c$Eg_|2Ɂͥ!\iWhRiZrj39b X$vR"c ;Y=F˔#%)mx) ZZhJthvk|A-_ǀC1GE&Ⱦ}bP!w;C+/~? 󞐀j+หQE@ "rK~B}ecjN7/6 w;_ød/XQg.}4qߦֽ#u @QF׻#%*ݼTXWu?**QŔz޾i Ϫq K ]8XP)C\k1)J8qke@!4S4 o-W6M ~{2  |+`S~(*ߎ|"4DQ; DP inu<&K1Z}P1UF}e4\0u >A{}?_$Iۍ66ɣ`#ֲ iқ@fDuZkBidkXII:VXI6DJ*XӤh0KBQ_$Σ.yF!GJG:6W*Fp.(DڍejDr S-SQ2>w%N_OWh՜Y1RsDФgl5pf%wSvjfpd}pڍlxb]'<**ĺF UL^ F h AM)mN77"lJ0"Sϙ6/JD 5L5)T1 eu}_u~4+dV1cBQR[o hĶ_P}?Q@Eqx􉒤9 @+rnߢ:Qs%11n[Z76{(kXRJY6:YYV*KP-+]K,b܈ƭZOGQ]фmdΤz5έ.%^aƆ=GP:Xz4z( 0oϽ*JU+`7B|ӟ /1&C3Nj9?!PAST{7MD<2PX2TmK]X8lGxj 42fC!pd&Mjšk֢20$E0a3VzM=q<4IF2;(8m<<韓"䄃3}КѾ1ƤwQ$'5AWt_RgSh?`e]M%@UgEP=<93֬\*@ NZk<2圖$)k:2K,Qʾd@-e_9!u_kSvަo\JaL $,ZɦէI%)t2V>±^ 5%xj/h_",]ksܝuBÔ<5:p*Tp C:KȖnxb&.= ]6/+Fqr$JJ5)ugqZ)at ᄠô3UNows}e77 KMN98KL_EZPx/J}ox"fMPg`f7!qc3A( CK9ȸe'qY& _w]<Q)uFβ7_ ;%dhrJ~/}ۍQ,dmx6˾;S7$7w=ѠB6V4icNL2[e%&&K #ɝ0G]J} uEC3E vH[]Ծ|<86 A;} Wz* CC_uFPb !A%_#/~lY*D"[puKFbqh!0m10ekv"~oXo MDނ412M J9;yiU, R5|2pۉ\6Q4:kh5y8Ja}3ZBJNa.˨U8V8N,2% *B1'r#8.cZi>ƌ<;Nc\wtrU`ES#592j.ѿ/4};Q%?]z=/Wz`H>S}O#dg'ZtgtpIU|GZI|m(oA77"H#2|iS'` c'q*T_1B?FG&8^dH5preRyxn&YfA{T&)σUKyf f-fm?qSv\% ;yi\sZBNdC8a;-QW^G)ǯK懀6 ATAU';yIqHaCmʡڃ 䜓a.IZ29 cDf8@@gA̒L")hmw?%-orA+.C̜{ RIZtz5ȶ\.t||l~Rcԙ6aUML\?4_'VwAv f왯O.Ҏ96~8!>-?cfoxÒg$uWJzֺ|IŢ,+Zɝ'RY%DАBtFt-m< κZRM"iI4&>75nj@q&~x‚s-h/Ь_a^|3/>:0ū? _(?Rكb oXI= gP}* !LujYĦQ9#m;CF%zco1DfwfypӴt /mfZm(ߺw^'%v>.G ϾkoTm1-*s0\}>BGl9%&1&:M\hֽ;g8V]í~I=͵UfAJ3ӚvWmz'ej7;%c>[@t~@QQq,ʵ8r32<8%#0~;3y:gUh IlsY!W`.i;:$@o._-eYTWWlBh6`/@׃ '%v +[`|"lqc5˗0}};~rd2Hu+ PdK\a'!&4)q0.,[“R|ᕜSŸZ47׫_5XT_W?%WWq=db4Yad%o?~}5ȣtܲ{`8{ܻHyXbxL {h+ #W^ ]-rپ{:5 H3*HxZ l9,I6KLY*DUY&+;EeDa#| h:EghS=I7؎/U5Uv U5T=P ԇӸط v ݁\Chݧ^V [X? {of g 7Bo.l/*~`s\iE;OӳRګvcxy F^:$*Yǯ˫j˫JC,h+Gw`^zg2Ӭ[g.Tm]`7ř1iiWi5 T⩖;%bc߹d:ݴ_.׭fڭOx Qnz/+ܭQ[DKdҬTBW[/)" uXԞ:TԢU8z88 .)|8Ś\+hQJ[!< >!Ul+UlyE2 %Dq"_l_ Q  *C4Z^VERk#0;ǔxbJ'2ʊAY1!Byz-Mvϩ*F!p壕*0SҘ Db OEI4 }R>`Iz:*U;kMhT$m Zi1BV$͚G4p)o!Y4TLK"s&4d06L2xv'u& ܸ*-0i`h)wv+ň\.iJ54)YL#k 몏ԠHmቇy 4ZM\5(&ҌP-vC5iK~TutF/8R1!D/K x%"SU$S]dw Gt:_zݯ+oЂ{; SfST&7#̼Rbr看'_(W顅D߸TKIƔ X"QYAGfN@D[-ٛ\ &͹zUy8gdrh  (2'z}*=\wYV{RVjYV ZLe;v߳VSW[=a[՝rv爩0L:0'Ewڄw_gpԛ_~(pqΐ+PήTewj֝8O4cɽoU-v~8s0We!v022(%2(RCACUt0̸d۽TTzB#x o{~2_Q!jjuXV:]9diCZ vJHv)-^mt [ϴ.@YmBiٿj!g%o9yrON`gM(Iw^H"iڻLoRr~^,68㫩vB~);D1 +RaZCY6F"Cf)r=H2P^0RfRzcmvi8ڕr|FbYQ\ ]`Jrc & dLHF-8mEWxvO_ d&ft1 ]@+|l 4cL[/'W]<,@8>NIH\֎FK,M914фr# i?kmh+ ΢ 3Sj6a6Sdٺ]o`׈FIѻ:*Ęhf|o&. 3,WsW[z9LFÇۛ^LL^ X[e85pK"וA1%e>Y\H=8 sZ##d3.D);nb7KL/4KZw} 4ZK f<ϠlǶ߻(p]3)IlkL!PȱO]C){-ȅ2Vο)Ԍla;kzE N!o)fDҧa ђG_ e ~CΤ'['Fd*1)*HuFNrޛF+%$3hU@MQiB4E)-I (bdGksyN@dS|W &?)FqKG +x)d3-D=`#U5ϛϊ'zS{!O/(ZH+͈ dfdFA =j|4v]B;SMf!yj4d,3"c6N V0 G0O!M*U9VHjw#l:TJ8sH.a4&Sn}/`oISٸz\%q֐[u'꩎QT 5;|/n4N1JU8-gR!d 0F"mw+5XGm65J')"۠"ҿ ": |Y/V.k +xu*Ğ'Y{*ǒ/$לUYhKV] [ih0!M[%UP3*F-I^T(-OԠD //x;r^2Z2Ry 2 Qmvm 7μ{gZ{H_!nrÀkPNv)jHΐ=o2"JpTuuUjc;tgH>Kr4lX68]39LnMٮ 7e%hKm~W&3˃UtG tspV0eЬJ6^Ѭ}݆ȅ/= _*N>݇ObNL7T|r]\*j sqՓb4GIdRS+12zue d%9~WGx3f-e@ѧ2ʉ Ԩ# ߼.Î!S$XM zKTϬQ^HvJӹvԸמ*z}%̦FϮ*逸>yD|:$G/H_Dl}Q7'[ 1:GrCJAs֫Dޤ[mkMQ~BS*) NNi~[_;m=Ie9P2z;;3w`<.=b? ߕ^z3-H* iY熨RRp Onvg) Fa+BE[:u ܻsIеav۽s е8θ@Rng仯ZM+o+C$8:`5uC L..q S.q;GhwFXNĿB N;U  Cz?3ς/x+Bt+? S]Ms?dfm0?pvש0ku;u1ag{5xHm4nw*לFc0P qJ r.| GW $'Rnȝn*۱oP@PC72w"ڂ@}te*ԗ[_Xk^Viwfmd4.^%3YB*HKXZG\,ͣ*2e͠zUd4e s'Hд^mq+hWC@$rbe2}7'`:{OﵾpfGE\]?x_}YA|^]!‱" Z4*bbCIxedT A٫1Zmh>[Da']KVdn2zbʥhޖ 0+3QE6u~1v92>Q# 7!23Qrg TLqBvto@)euWC\ʫA9tG_ǒ?g:AZ*x+K|1}h):Vdiԍo@{ᗁ HKʸr`d"-;w{+28)/a>渭p+ӎHܗ/T+d[l5F< Hyfgg+ ܸ_ %`|{\I%\ J@ +*QF\4KZQTiD`S*D\a?޼t$SA=i`"b P",Ip"!v+J yd"cn-J6*!VNNegDZC"j`@k@y#Bq# }3W(p?Wz^Ѻ=u;^o{XXͬ2bZ=hΘ)1cđӪVCh (F ͯڞf2P"j锪6Wq֘ .0n?{ QJLwb,e)N:Jg~w.^Q:(&d1Z<#b7I0cujZ[st(Ɯ-Os dIc 5[XcYDs86}@Pi]6$".]u+f~ chӯJ}mv4@hphB ;Q+C 6sf]χCC͗*hH128腝?c+fɃzhsސoew|.aP Vykuj/KUvm(YDs ug]Y'2܈>jA9awo~  [y!#o4%X8!X)YZnT 3S1J˵P:&DaQ"NI-$rc"pObh-1 BB)`aPI)q+w9ь=a8Smk4Hʆ#[_|1Xó8q']`Kʇg:@ 0)eںm]CaS]Ϧ,-EQVBGJ;FGI^"&aAf֩vN º' 9f]rp˺v4?vE ޱ꧈ EQ;z;j`Na՚ yX;yXaˣkVE<amaQKyw gRU/FS e4ʳZhɢPϟwRE$}< GKGh[v:Vo ip7d8½K{P ]6$\L2ngw b1i79mbݧaD㚐HO@ NT;\H'*AsD`L`2Kd%Nv_-@9Ԧ ~*cbP"lt+M'f̚>NCR2dRHd̡х9s;0j@>tPMUdVrNu#g't=LI^j~FgfN5mn&W%p2![0܉5׫?7 l2$)eD??yEԞ]$Wѱ3;}bAK8;2#|{6$d3LX+ΪE:X?jP h8r¬rE򝑚 = ܤ~ØW<skRFfjYAqSƉ\,2dulJ2+eUj존jNɔ;Ђf }A>M;҂`kL!mBV; vfphk8iж:9}Nx͵i$EpNWj/V_k٬KyP!Ǒs|Ό:g99c9 0 iG&dk擭77XKB:#g {<~Sa ȫ75gwb5q G8 8jfy?P8s;]Gk\+%azxۯ-gfr_hឌ¦9B[~2^3*I9yx=YdSo/mU_9) 齺6YW`\ZoqL7߷|OƁ#0U;Q0=&4 z5)7I,ȍscA~B|ux2_:fl;;1.}1=)'$[FkA (܌m4|~s68^3o5TGvOնUA G VF}x4 ]@t8NP "6&;=/4AOm2CENł#x˧W!0;t z84$#UGGTϱxz$,·n<~so凭s|KybR0,)FZa[ +l3Z^a ĒE[5 zL$NIU`VZugt:G Q#+ĺ:Yնwn;I}2Q`jN̳>ITԗP@d|[c/vF4U֣ %C4`&s5;/W S_Č˵7Oڏaj6ލ.SD{Q)}&1Д̨ya9dC*gNc ] ~2Z d_l)gݩTFO:XcsLEJf:?Nlýo4੽[2cH6Ayw8v&X'C*l =8B!$&I[,[v??0CE.SȈ62ӆQ#4VVzO>^Bs ~oofd;!!vnR bHegm"?1ĖtHJLF4XegW \){\,G0jqhyj-c_>o(k7fx#7Gu>' .wY'%wGAg9J3 FPBmE^PdKst72 s:C Ղv} 01mDG`X+މˌNeFGlW={~T;TK {Y\Mm}r(ԭ5{ԏNɡ7CD/>yVnѧ?:OiPCQg5{^BoaofazLPV@}$?WlNʇzLI wz}@n8CVZ\!h!( ]'cɃcB4mZbB3 GH&!rcHMPƎSg>&ns̑ĭQlsŔpm0r )F44rGՁB.ٝ /$ l"w6=TҺ" GD Ƙj(X.UJ5@KdIYŖb#(X HZb2)k"%d>,!7u%Rh9A`sK@RB)L*fXPP%5d-By `&Ҭ6Y圅q0>ޗRJգ^|RӛP߬^}wŠ7w>yLXDG&)Zs%믃\AP23( 8ѱrg Tպ*Z"@BU&t!~hH[KG@ S@]{lRDq"a Y U:YEXVi%Ą7WTڗ ^%nZ:lq#cwZNh !7Ilh2P}U8 "ȁ|8WE Z+~=>NC.T"f:| N!;3e{ebX$ W5 >TpTUpVnpٶ{󹸵~uᎈ[gӋ! +q;-w2X>:A RԚv'_;$$Y" $)IQ8Fͪ@4f^q`DzUthmitDϾrs `a$Ej/RxeIdZ܂>yJ@cjǣ9ʍa8^j-"?vM=*b2Gw@@9?r89\Z&ÄAU0g)HUVD2c "a-\A!;&wcg?S`/d!S/FsrV`cF"vf(`΢ԡW/=*,b Hr-q˅FP@ي.KI DG@G] SH+9Zj)TKD*RKxD׏myo3;:zz/NmMu?=p<3{er*%9]x",=jH-ݺAbtSŜUD <;4JL4A׳16wZHai5p/1 Tr-2&Մ.kv8:3N&(Dw:=+$l1冱8음}3Jљh 9,iB];46q- {Qvҧ#S)0e 're3TcPopm1 MJjq-ErUIdC|NWG1bnObۙqb.y%k4^.b7Kmw,)z'n ?/ټ3) SV)cϔ}_*24yT|1U?JFne.ҳ%9f2>HBs):5nZ3-щ}GvL穄i-|"#S'Hv -щ}Gvi,}-YvK!!߹FwvΨb1wn猑PPiڴv?'RHw.12cj>FhRإltޟO\hjY_ҕR9d%{?WEXBDڅEl:|y kd2cȈ&s) B*VaT  T;}*p{PL$=; 8-`0mw_E {[wqa{ov{x) PLjJd~w'EӦV0fNsGΤ= ]B^uLXeoCG[ ^l~/2`4kj؂7wWpJ3) &>M}QNHj %F6!xk6~ B*݇$;Մ9@3j>7]Hy>p^~^}XW9ܢ~21g幠1.uEfsx3+/)Lb )Mu dx"jl1/%,j0Ys7+fnn F_kH+A^O4QwmH_eiG`qHb<>dۚȒO'WlIveQbw=`ǶZzXus`t2{],$f2Q`+X|U(g,W#I՗H`Z!u,/۴\>"$zUw9о|Y(n^_\ݛp_- J_->}phJAva`G+0QBC3>s0 2 CIit+//l99 ['QOC-&Q(AݔN8Gս0H<A;W{? ͹F]vp" 617]?78G3F "*raŸٞ1V$C !fgE`9^F-2d[sǛ?SD$mjse/*YU=Jv%Vߠ'JMŦ$#Muiί̔~Z93h٨L9(ύ5DC!tSS^ XHΧQ&$~[S yƼBGR cAuc1>ͱ|)`V9(E? H5 놳_,\Tim N(C=phs ZԘZ%: k!8텑N%a&`T*Y$םD^M.iZU/Hhٹ}ŵШdܞķ?3uH״{fk)،JR[߱Y R#g{  ]Gx.IS-DQK'#0'^ CK>%1D\o39JpjTN F@! Hu"d{ߚo|Иw/NNIҁK6i^TQ cwpE␢;ߚE 0ҵ_+.>|+1P2b9΄L#gy&q{XU̠x՜(os1lT_fɜT\_M-xp_N &Qg\mqHaUܫWL+{TE#gB/rT$<ߚ,3G#L"߳דl0XV\Y⻋OgLf y 34S1 qVe 츹Cr kc@/jx>^~Y L fhb;ߎ}o5X8XVa~up`Sc+d IQU:ԋ `?(n_H%'r!]>GA xk(n.dQmaqDChOs.Jq_ݰָt<Z}~Ab`NZN]t:/ky/x5*?pޯZ߅.ڡ#s}3nP_ z Zh, CWjb?~PZTExyd7ϸ5Gb鞖{)mfo ++#ɨuC~h:fUaWk JÚ]g9Ц+#uQp視pFZ{|(WPƓɸ4}z?ԁ^>D!l> }95&N"vx#ɓ1/DODoDuEL)ŭIO }(iusfw.GnY̕vZF %w؂[|i osJ e8n.n6 zE@#9$6Oq:Z9A$b'bp1-6vS'/6xp (&X|~lSR{VRFQoz+OI78}1D4fpyzumEA˺^p$k,Z;yTc ǕjDa ?Pj)vTb$ q"מX)rx^XW)Pn,H">'t[®e,D)?Y>h ՜ny"'{-,OΟs\`]c}Oeu,d%EJ=H/̞OI cb8,@jQp(bͩʙKoEAiPLj$S3L5D(Z16FU}`|j;9ɂ';s@!KWe  Tx.61A`f.8vqb\{pu-E9aހހ).YB([xI Y*cCRxa`r̙rt]%¦e >~(|/G[~a'II%!3Lf~h>g`2JyS?$鞛/ >/wŇ~c-8%H?|{TAoۻ~Uo㛋^1Lg ~>F׻ J۽>Ka=_f:Dz6 )SIyɹX /z%?G#xWخQ[SΆ}KWza) MJEY}>^G/,߂@x C}u3 95gy&_*zzLjk0MY, 19XN,{\,_$?ŲpEhr7?ϖ髫+. 0]b_)3ӛ|l&g7`݀wG+ .ٻK"#og?z_̧R+4bl9^*y@=*b2h$)Ȗ8P\YVo#9W0-HktW|}'7zD5Av"@"D~j5H@_5N'+̛Y%@N} #~x79_6iH:cL`9}43xX. 9# um5ǘZWx#FB{G5K.QZ̃+@:9Ʃ%~NmNJ6w+a|bF#mzu;!ە q7T ߥ[!xFW.*}{{1: 9Cu6yO&"DɆ/ՋYąDT_"Q%RѷO/Z^,ͣ piklP[=x:YR" Ck8ac瘊:ݿuA`P-+SģpߧVR1v n1 B{R0LAZp-03k)T;/%`:Sj;;մ_zBxrwS^xU:,o{}quo&WA~*lXW˱&ϘDz)Yn&* \ݼ_s1Ӈn){ԯNe6u !eqWr4mVӛ?}𳞙NףfLX 9%b:u)Q ƏV'vhrxƈJn1eNm ۢ NcEo3T[Hn Riw/]ʛTe+YE[4r(ۍT3u&gJrfV)$O͕hՑ=UB>Y_PGSdl%PHiZ1tJz S>?{WƑ ݙ;4lذlî!)$($G-Ŭ/"ȈƗ_iwp~хˑivq>'y}TI6LOy\dd>Q~Cp+f`[ʘz/Y=)P" Xts3*nC0s1Ďoq/ssC 00jY `Ec/$.N+Lcb4a!72"GU+nc bLq!ҧI†v%A/'ms<W8B;T23뷮"ꅃ lE?[^&a%Y -+e.jMVքn_Z@l_9~o,%dii|dO4{_va{ʶh?*= nmoeg7 *ժmgy(bg]Tk^-bc{t>ٺ<+?;%$>c/έP -}B=Bt9HN,vp-"Qԑl4Ja]-rPPmc 8Cm,Ǿ[vWMCY:5~@@.|'IA  CD C&sCdmC`d%([6Wɤ0;ڐ̇=6"E: v#:bU3L & d{HAۘ/ߔ YVPN?zo~l. x/@[@lQ܅2[d~H–vvVu,[ֿx\0]ʤdꤎ~橁7|'|oUyAz~p8 ] ̨]zT$@ !( (}!]ȊC/+Ͷ 2 Fo{QW1N !d}zh{7W-h ٲ=vBo;&%#Fʦ* sU;p/FǻtCȹ nC@fCu\CgO vPI W;@m;v[9tוm8NEh "vi9X+[({C;dˏ6\W.iY2cG͢w̯^NqŽvDjI;Flji4y#jh,NZi1u~As\D{Z[U>q6Oc󭪅Rva؞ЫW||${x@K8ex`=ҹn aݠAW/E=7)W&\=d..H7?w][Xã*ڽJ+zM5aMO?s}bjt`fJ÷Wxo=ŝ|<=*8(ڛX}rxXxPK| _n*C%71:Uv?" ?yGXK2\kiAPq&00>2&H J?×^tryte- _SVT5wF545Ȉt2HTBوCSQV(V R }"FfQ!i jIQZ[ pT!8%Z(AGQYƃv2L2*Pk dFcQ; 4x 3ADK%7Ez1ܨ.1#]n ]S j)7}17_ļD᪲ BGTP] ^WEQFG[&uC֡Z/iFp:<"NK=ԁrg'q0o/?Vo7WɴM,&{w"zn {m_z&ɔ>|>_Fd7\O'gvvz+VWU&d7KDӛ(\]ӓM߁#ɑքK0ƌs3뛿\Z4#0͏ԑD d{=?DZš˒ANrG 06}7Ń>4((M:\ON/Pd]4[ MȊ{++4&C$dzD$Pph8|r#E0gR8dh svE5T\ljK<茡SYp .0 % etL^Pmu&om ADpזCfP娀R=jfOV9|oOIZzl^@M= ԠƑZFY0I^jss>L'4XJѮ?k(IΙX$ 1ݷ~z7]_^y4q#M6-a%A}?_2EiH-QԆjE:3o޾$=O[_~8L qۣ~Av>L,ʴ|=3ROq',Mc*f#j듏gf;N Mɮc'RhFHV}qGLZM?m#vd_sf̴ZT׵?f{V 3T0SM*UXea?"Z=bq1J-M4YG=.+ZhXHpw %";WH"A6-^ :P֩%iB\$!$Rt0]u!`hA]2ʸ8]{#fkA \q+=i^U}3D]j}cI(cd [K%%V 7=H=If JŧW}k]+.yEFTKs)Տ?V.^ĄtiŸ <?r# .c{D.KLI(ru'Dt\ 1j꓂$$T4~:2O:93)ã/z5yף| 9/mޞ䝣O'萜/_C `5;[piy2!+Sg#`8bRIJJ]|%UN>B=pG`Jd$3Gъb6 "C)9넝Uy[sP̩F>Ն'b!EfK*)ThBSMt-!YI reh壣U -GSv䰑Mfc OVcc J5mЖFV_xff` nlȎ\Q[x3IV^zxCAn(a+# B;J s#"ԇeRoN]W&΋pNjx@A iH[I:>=q>gM%#߿ѧ=/Hŗy.˪,Z|/m7WO)֠j(;%GsȮgD~Jhm.1*)JY9F(F^gY 80ZmYID۪6ioIkzZ . /pR dSZN^x i(^ r% xDI$c-'Ztפޯ3J5IƼv\z"HOB[ǐN$m"EAuAzAVܿ/%`4X]p0g)&D'Ngy;y'ǩ$kJs(%i`8B9&{OB$m ),kG0catδnh9UU=M?%i8r6K^K,{/.@z׾Y I-eȕCw!Ȳ oIRwߒ3XԗIV"Aս[Њ}-VnnCsu87Oc?+|va:3VU`쮘Ѓ\M ]&oj~@_%rB4ߴ;r5Gf3?^ILdUxKs#w?nH_KpB4ڋ;^lvl`IX<|Ɖ~Ei,1b73P_Ud=]Hg\7N z@'q('+Dn-kDRB'Gg4 O9n;\.OAR˜)A! kcyFHт4sj'Tf(J))̥N L#›бuLy"!mo'u‹4uğȝ]sDV0@_j'kkq;`h <Ԛz\6ZxfAS;H,1p9 *ٳ˙iRG^:HU?y6;~'6Ö`$KF1Sp3)B}oi8>љBNdSKu&/ O6 2_-yR>Cp+Id[G(Ԓ2xRrJϙY JDxS0hjHY..߇+ GfD5u'(V5ƘR(9mx@PL-vMc]fvILȹ0 g?7C$G1f)1s*EɜlSQOj8H`FjTc8xBB ̍,L+w¤#f!JhYw뾡&nřy!PU0ᦳU0gXU͕ٚ WOb3O$O !Ͷ:-vh޻)XԃDQ_oHp8D z5ЦB_WM >Jظkms5jEhUWR2cWSbRG`CRAXg39ϭ[gĥv8b-cpcyxeGQYQj\D](Ms\*#f.grt˥PV^ueJn566 q?TJYm5WRHIҠrdqfPϞ]䴄qCeg) m+`ƶXpede4S ET*S:C%j9hv(=[6 :C^]ϭ WV%5ؾ)z1L}(G\v˯v\h{ V!04)T-&Ý&s݋ľ[ 90/S9/@ݵ[695Ɋh _gPRA`dMhi/3慃av|班\~VV(/T[􉪀KThU[]Q#P?;-+eрZO>Q#H/oOr3Ό Ih凉]?rԀ3UģԳ}+C'tɣ"?n:wm:w$ >zs )&dawAr|;@s\\뗅E twY.dQqYiQRGyy`hqFP%?av}y_mx#c 7ɭp;l;ᖄkNMng$G^ЂOLk.evqs,SBң-k=SwhJ˩ԎDH2.i&WQ PrBaA UP)$9 ,4'KK`d5թ轆+?XqggL~a{Z%PR֙p;{l?4n%'imyQ] :׍Zq on:Um~iX̮nWg +dHjO;{ˁ>|#/ 0Lg`5<ɓNh]QviL!L$%ik#S!p]]LvuBДAqeqzyjTۜe*XBEiKH檾3M }V:&VQ?O,zɨҵG7^sTiˢLeE4a%+/ BMsN=1T"5‘&@\2a51̢Q;{gb &0.sZ@v>sƥ6MQǚPxgmD35*@ʟqxd@zz:ЪYV ՟bIz+=)\W1@y/VTX|번L݅c|JS(Fo3;Fru{ ~XvըN6`|V{ȷuteO܏!+.&6't-.> \#;5L%y0S;T.="UGi1MђKcG:^RXb2J|6᩸~r/=r‚y6]YhˊSv:ˎ8˗3c }\6wٿ0$!\4(Ő/ۉNdQz*@e30ո)t$lM"fʌb.ljia )88 ќz$5Vji-aX68,t .%m{>ܽ fé[' rh\&񪐽h\Jh(F9)0s8m6];m2h7D1Nih2r) B G%' %ͦicRNɪ;湄JCA mPNI^ꊪ r>?M?t&0a,p#O?C]Xc vp$S.aJRU彵T'iD0f]ZZ@=[ K4si$FďhA-JP%`Jē!p>Wm1 n(Ew쫫9ز[^vHs/oJRyL/:t1Szc4cyAN=k2zӆZȆ9lː6'D9Hm ҂C(Gš]xq* ܊ٸC竒UJ5ɰ~շo, :59E9$8ko G|0VJ^qUTԴqEy&ےW۳ιfrCugq吅knPsԽޣN7N {(|}oމv4(Bh#V(ƨF%PDp1mh8>%Pf܍W.JiE/؄{b&H=nydziqT fUB5Z.Hq%2q?PIxy:v\|r zW}V9;]8~؏r6w`I:[GÊ:0x H55:U #t+Qqkށg5>]J¦r@G 0?rB_T\غ8+ p'5 )eV!ZJsZdB}ٳ˘G w+)JJK vK'K;F FF{#)lXi#[a5 :[i6;bh #HcDd}$1FL9Ų>~]%\" #3 1CNC\œ 'MbcƹYViz4 ;1? 1?i8 8!?[h 0nmLڎ(έ- EL߭@0]A(&?gS -^ pH6 NN].dۂFvooduGB5:&0ZC%Pf:_h`@D m&'= Qlc8Di憌bT⡤C: \ﵔYJdnk̛u-sFWڂp,o#YB/ّj@O'bg,MD]?$%$^z83P!X6gSUuݺ{N*WTy` I(-Mv?JKTYhs~v_ӱM;gS!7(IDw\go{fp 9;w;t;^40Y(3ӻA9O qr8|vWi1f?\EZwz;e!8U×WgS(kv2=Kdg5k 2yEYQ󽀿?|s#uҢ1Y6z3L(Rmi:JG&̕]O&s ΐ[B%;=0GT9ۑ/V,­flP{#I_<2N -Q[v7nWWߤV&rXE+Btōf[;E852VP)UT}SR@h!B(b7*^x&2[[]Rw3 L3_ ;]1ȂM5do 2_bŠg1b,[L+S/[ā!E-iKwp)vԫ3O"iDolm/W)pJ(W+t_ƥD|1:MIg2| ȟ@i/hk]=l>^fw7JѡC mHz֣M{1]}qں:O@rqL)acyl+OA S)6;]$]$]$]TYWmk,ҜPF Sy)1E5V)cyP::ඌ,ql5j[2Z=r|d:[?i.Ҕ϶L),к+lڻhP 7_fN jq u[Yyuc_N|7Cf7I.Y>BvJN Tɩx2g: qNgD5BKP C҄R&RsE80RQ5 DY%#th#ghL)EȖv6en/d) P'Q\N?35TVm?R9 h99pծC ttnEfÔu.ժ`X4,6j1ƓYZ *sؒ{BԡqH^fQRNX|˯5ΜbЊg:4Q, :aqEi )x0ҕ@!3.V5g'5ﳦ}ڒ^*F;5<֓%!{-Bm?G_3gHįxFVIq ENr'v{j%@:KܓJ!g+J|b56ҬsJ) lնC!d[`vi{#QUj0诵=PM:^':OQa< (S]i9a$a} ܆hh(8A"PAJsF"96; qqE{T!KWJo"ּ7Gh5XT!ĺPtjh/VjAᘏZQh ;1 R+u=fr>/Q'oO r j<N08gds}(.=NZG˴8Ü$a$Ht?[ bZ8\d0}v2 IZ(&g'k@"-؆Vs5I2.DBRByf)!0D8Q H) Wm΁&1z9u?96H-EPK*D( +hDi4S [ RiTԎxRLiYXgy!Y e>OρOht m*p(C"DL{4uA6oՒqeɻE&Y"8?8Z$L G18@IUPs'UHs@'5 v: )-@ڸ^0'C$Ѥ4%LBd$Z*JԢ!Z 2̕55}aӷ7ZcH8ba \݃p;J7xqp3*ȔT'!ޡM> ЫW7g'0.Ex9VQoޝ5Lr{ysaw)F5N<%-N? Gɗ-j(M:@?cr Z0z鯷^lxIQ0Ml<[ 'sCw.K0?m(hyvVdz2RάG%aU -厣<? QPᴔJ蛹"Zl^#i4 BW"HgzA"ZME*mJ8'XA"rᤍjI`e Tzt6*-tJPC* b3/~Y6|'_5)IZb\Ck{ H*T5La_>JMHҞZXlz(r( x Έ7s"z6R$ #bp͜sXo3EQ Vkj@ϒT{Hn }3gy+YE#:~ܠnӀ!%91M ȗKǟH@CJV1 5&gcsSw炧8w;hoj j;qt# 2/C1eIyt$:GXZcaB L!i`LEzWLTqSogbf <; ଞXqɠ@)I:@c͌\.A6jH$#bBԔ |P5%az^?y@ $:e2vq֠zuF@^ZG3A 1hUK["+SʕS\r|PvL8|Z!i%\ :@r(8\m=U2S,yYZ҉D0 Ć=T~]@S54\ e>bigUY ti!:~M6?SFk;Pʇiϟ&9+♎;;4=c_.o?G,Aoh?-لZ5G+Xg笫>|/*駻Kn=W_+֝n f V=0斑Rƴ|+Iƒq Gh08DWfv-L(?Z߻NL`!1Ў'~tC~X016CX&7 ~aVi T%V[ݪbߜ@'X{'WA`μ~Ps;C~r*Crfv,YIEJWDH 6VT 'DeqX5wƖN^q>WܐbR"E=~Hj(edx(lφ4|@$H+HCĖ#ߚq!Q@״1\1uD:c?qgkOucğm#?#O4t v\:?y{j1kT0؄v؟Z6ЖY M[-|]7<*Ąqz9B\ٲ19cvB+k0ZB./tܗwUd?m,0L\X] c!(jԚKC`,|/!i#Wt`ǬfY wõ̇KV]k|9~-iiOz:^|M0P$9nb+H`O%࣬OܷϦ+/W$NBQTx5Vp;&H@غ({h1.hk8MU6q%'y-c$rOċRHMסRr\-8pQ;ÊДT wJLd%?RfY#dw7rYG]Mo)Y=mᵄ/~k/s6ov% @P֧@GX['2oKQ~ REIP\v@FNb&&çeR8+3 Xk atf SQD8u(3™p0̩"uw4Sq9T:Nʗ/VO_u)@V]o:JN*ah+T0,)iiTҟ-TI^|dYJfB1_iªZNI卣o\[kǘ[ѷd,Uck@幅LQzuMD%+Ȭal;3Zg~is9hGȉ62ZX<8{]eJ@(itLA ]&܀Siy4?)~Q֐_޿Ivn6XTju<6pp{TDcK(NQrGc(,Ɇu!sŘ8* 0_3T,r#hFKDd[HN3n3aʱs 2Cr!t<0o^L>biw[,"p|~e@Q b x5!pBsphPMtn)Fwşw*lN\.=BPJk(¼(Jk_!93%SԮ$^XqUENw@YE);u:4}Ϸ1x#/?H dΠ|-yAz4: 6 9.7x_Mˆ}aUcXxῡe[4=-y.Ő<Jw:~|9|:qFv8+1ZDn*|.vmӰ%}Txq-hkeWxQQ*h'x.W/sK^/Ko_p~ 7A IQ݄ߖ~7fGׇooȿp_|GW=)"ԿoSFt9@W' Xs@aL}nkؤI#FN"rhMm5TR׽OwPH lw'#|yC{4i%ki>UsqUsqUSK!΄ϊ{^f] $%bD-NZu {8![]CݗauCp$%|p7 :{z) z6$ipm&-˕q i#55V :~fdR3 v3$z2Y3ʘr'v ')hVXC-puԡ8c"6c+I{51I"$7 h"cvƼyLX);+:tg"9ZQr7=-HHf֒Wp0>Bb %5*!4(KX7γ¿fK= %CoC4@zèbtyU@LH _L"PȎ[`~fvr>Ra7MOw'1 ϸ !$%BH㇄F#fa\ Qo_9dL֬c^iQVF"aJTcq'bJ7c{fʚHOY3D+fRKͺQ|0z9gó zCݴ\7<;7- 6 3&ڻ߸ou:q!ICMX ŝE54$}?}_=ݻhD}\/-]=Y-4)ǃE=#lÃG{2=bN\VTX $R᮸q (mxRBa͙giYr 7wX^7 pQupzuRip-lח'@P8皣Si/$:ń=;4:_"b"1a7X*JVʉ4YN2ceIL}? @ɐ 51Q!q@jRimityqU떡-_?R| {ϟ7b' ZUEør4UpůPu21RuzE ~`}>526P9"Z^m$k0j)Vܼ%LښEiBA~m'YHV΀?tr8"3#A7҄,_5h^%o>ՌY'xy:̮1oMDP6oL ( ZG oBE26?r]6DOGKǫWnV֑w4*4|UQVECĖ.7E"!C^H_ i]r<:' &_D>c.>\O-ç .Z=7bDt(,Ӄ;tBrN0_K <\/@a%ڍtp7?\c,^_Ǚ],x-?ova~va~_ށ`RYg8Q9Ab8S9(pƍ;gf1pB~P-sq wߍ>bImnH8=@-_h-Z[E%4ɓ?hJx;5 P_9&oA)go7JQۅvQ8&Ih1*, -N -8XL֚G*)Nc(":LFLDKyeDN8 %[YGg\=ݻ__-F[Yh"P1 L6֙ggҺqD ʚ֧ZΚ@[ү4e:[j (՜[^ =!#9xӁԪm1.Xzk:(@?{WG` Sgأӳ;/30xe[vK>v}ɪ̬,K]Yd0HF|q}_-tg+Ro| k|T{5?}Zh6XN旇fAYSOx$;\V;ujs}ݞjPRX.}=zⲥuF UcJWFZ9XZ JNm}=S'njm‡h͘"4v"'p:?L0u*i4& 8 .d N?ihyWVT$z|ХKzP-/ڤYsy} nW0ޝg*|ޠOY}\V)}L=|wdim,R\L>lInTP2@HN"AjHb "6,ZAFI'{sP.\*zF4p5ą( .CpyP[)qG-Ud,HύBLzL+ 09P57kŭ,9^HXøtl,nbuyGҬ&*f+|ٌ`I?Ehc&zZ[}M`0ʸUIiϒUmǒx@%QxIL(.I%ҎO粐}Djd&sQ!ӂ͈_~D'[wv{aL',?(9Cf_ۉY1 D 騹Gg5wY ORÓPsPcF87'W~1`Q8<^@q8"Ǣ8BaZiPAS*m`"eP Ҵ\uRy@@e$r!8 `p sBCTU@u%D(T7" g^Bjr2QkuedΐB:edXP*ZۓRKި.KcT~]7pekc~ia6=49bnIm8K[$Nc'Ӫ7߀4M8aI>G(h5xGcK~2L.4)#{%bz _Ky6O gf"KJ3A+.T>Q$>X!-vuQln@* ̅aѪ&$YHp@d:t`u^B#Csε%{<ҋt\lk9rWh5:n`!8 =fPQ ȡ .-Z4b0| K~[jvsŠ$,>Rn\9ZoX #G!/|> #n/aw+8x$B{&S|\B߭Çw^8kˁgfΦ[Zɗ# ՑHf\3FH+d|alZhsxsZOI9 R,r^6 )٠~VSƩ7N ᄼBy?5ԣR:8k[7pF<2[=$[ (1|" V`F3Uh鉶AUȡ}_EQu8ZK89@ՂlNs`G&BqY.@.t]tbd7tyΗy=7=ȂE)v]AFjtS$ZΒ,# b؎m M(x?qKZ۱=dထs츓o VGWW,0|AkE4E4j8uPr  W'.T_5X1o?}j񷾋Bi Ǖ7.w%Cw8_;Y wNdj [_xs{\G=^do_n/-.~qri f[ZAM~"dtyV2-ܥW_7ǯ|L_*0%?i~/LCsAi!4f)/LO,In1EnV!xy]_=gU} lb L?.W9=!,5sy+4V9ƨI:,.Lvs$ vw_MC\3Mx%y @ab\EH5Vy4R yw[ YⅯ8P$d`1V5t=U}2@o'ģmT%[U{dt PYF7u myT^aL򨓓`u h%Xp:2l #8\+FD:EFL!Lh<d0/:ڸD(yG+:>f88pOoHPַ JåYk1"QӮ>W @YdbPWFCIUtzת0;u‡cP08sФ@?dl$GEg=5iaz$͍;:&L R5venNl8dPg5pv*g b0`t(<9{-ɰI{\^l&6F!o~J١6hanbVӲ( ɒW}>7}͝7]&oц}^E_?uyX|.B9C*#([_(yCGG3 qb.K_r4T|aX "[8p0d$ CX:t-ijoJXGKF?Ҫw3휘x|_)?(n)n䩄2ŨKU<xZ0͍),dSjRf"EO+!q.C/I#3Ἦ][?ǻ')Yk5~&M/ n9fxW#/yD?X!4AB MGΩ}U>~Z^۫au7+[R>Eb]$ɕq7W@FE㣉bfx&+̜! i"_xdm^7WB`!OΥ(Z :hxH7r4$ɑwKI/ȝ\0֍`$xU0!D˝ՠ(p\5@Cʷ"oN,|l.ʱy6` qkoHh@S^%%2 hh%8PҜ`XK#דD%<Q(ll#E;Si؄b`d^6ۭ 5mH .sԛk hvN A̘smң9Ssi_*4Tk*^OH&$-r7 ? +cA$@:$(➒i"Ǵ:Uyf])U&6ےգ8{P Z4əI]Û/6%ǵo-x9o7k3$KKzzJ Pl}_v08 d{gRci/Z) rO <-R^|[ޗ*Li*d/♽H.!kqad\U)@B%N(2=Ibdf=l"%k墣 w.HM92A~/{!|9cx ặM^h^}a}/ { Tza{` @#X o?2[5&9tw<w<Z _mR@+ؤ/hc4iw8W D5KߠD6!!o\Dd*(H%hO5u+;k[{ 0c([6 XVoI(vz";#&$Nȓ sf|HLU)Ɂ94vó+,v֎VX(hٶe:LOӎ?Û J'7EAكi kD1<nj؅0d+g撲­ (]d ;'xbHCf]3'cN>C-aQI}8Q su4m60IxRs+S=si\+\SI@PZ#l!^RΉHknujL b* j@pSݨ=dI`MVPB&}J>tk~U͗%WsJThՏcmgs[:1,+PY,<»pj_zdP|zT<.62鞫D?>U1Z ]}xYz{O7ιSq[z||)?NyO` mTB}?շGqZ$%˳71+k^ 򺧽H"Tn{,qz8=.o7(]oA9A*mG~5bBsN =HcB kf=ƊPad>Ԏp,²|\y~3 q%6RM_Q(zVbO dA&7wP)]'c/ⱺM\8<ᤫ|'?dt7`]]trqArǗЏF #'tބ:'SP~wcILB![/0: 5Ím!L(be-\!T" 15XDPjLt!%h`֚ 0*VD  ;Q* UTRPvkk]C*Kn8BM(Zh,AǟCeBR2XJHE:PRr୥sH95OJU+}R1e9RQ! 3^X ڸ'()ks[U7RǓu 5eUTce7@QVV-U ИsVR BV 4=w)@HS{/' mp O2Iƒ1!Rx8<|dG^凯p$Eow=frkB jӥӣн M^A,\%FXjQ /e?u9ޙߛ*Tute ss7cU"aRY ƊX0' z.V/}Uwna8}[n- 1,Bcf!+@CuP)|L9-C5K ؀i-Tp2KJYX+s $s_>tS,gـLRi^Pe 1(H}y@|923bῒp|N1SΛE8ּ{}2&C|P#*WǫB+@EGZ_IW+G2sNRiQ&G{ BB&RJ5'j# f6,@cC4ΦZj@6ihaҙ*hp2،d."yO&4F6g/Ų>pH 9Vyj@ng/V ! FVJ;,tL-JA})p5w|NWG) ;YsC7[֜F=@%~wRI*'z[8&c<}@r.}@#i9tv|UѠp;Mv1r"QHn~Sޘ4 Xs7Gd?.svM'zJ~!6mD-Y|֘mx爆 q7JTR}&>$ c rP8DlkAIض, ҡCD:T;b^{UNҕ3Ƌ>[LܗxF-Ќ{ߦa~<~.>1ų-כ_jL`1u͞z'd HB5GN^<;#khYH bʪ0m=xb1.X ]wE9= zIh@8tI5BAse|^3ˇ24`#6nTVg>,Rr[?{Ggl'.&d ~={ w,TOyqggGO^h>qEu҅kM}9<+*)EQn!=*\Df7.)2C5[ 9[*1:Fv@xٮ[zDv!!o\D˔H\]>~8"(q܎2[v]mR [7d`e RͭJɋќY B"{}ko|<~֬n>Oo0Cп >92#~"o!?0%5X@օo>$r@;.B Ӱtn9f`&~37OW[\TM|qJ )zrcRdNUgnOy8"iM˰NK/R*1&( c.%뽁kpޘ֖8̡ Ĥv(K!Ofc/ٳo{иc/{$P#8c!3A#EjFa2kn$1Ҡ$' mD. oZ)agО5;\5ȝ :y/XA-G+ߠg&qsg+p =ٱ`br5 IC7eMa]N>CS 8[GyW~$ ԒY,]{YtRؒ`%0LEKM`}wZeQh(ג -iRJjI:|HH]*zf0~bKoܖ| 1 jI=\Lnr /3Q ˱ZOR_HMeu&/ږ o@5nj,T/LX/ LuVkt}5^ٔ|+0qoWj)zuE{ -&Ғ@w$վgMNg+G.QP׿hS8:q  [FxFBdkHs{+D~ƟDcBN(2Mߝ%,WcY/]R$zj{1xw֓zj2x力j/ssU/n-"V؝Z?;C-`67m-֓۸-zv S u;THO/j꫍f_{}v 9VwZ)86q犝"B2Tk> `Py݁1hP~Av=z4[ɻnij3t3P@g D=TSWN4 ɔ(8P]oGW}mÀ8b#NnEDiIʉ4|I33QĐ(9]zHB2hB)GḆQgx/l3zFX)F6h\E,7h $Pna嗅IxդfGoѬtcc ZJy%DHHYOD !He#"JZ n9AU-sg'oM& -foO5<%W~NYm㑏&tM뉿]]rڿ%% B J?@>`Y>O}\@DDhQGM~j}- =OqQ WΥYYY~z̏9O `Ba=?Sd?[G:P8?T ̘ TND}|.t^(j4)4dP-4ݕ;Z+ў,&%},Q涥MJ.s6iY57g9تJ\f4]qr8 *l;M)Ў` =^̕fDWy̩8*Ȝt]]Ƴ0]pJs9aS*77@:5:h5,l qo%0=}wTºzx^&-[xG7W(Fxwz):5 QY:Oy*Ei\ Iiy'hVlkBB^f=D3+'n>hy/\[~ڭ y"#SBvCYAbPGtrh_y*n5kj.$䅋hL-)?~z_% [.uD'&g$JVnOn5[EL)аFYD% ˱+u0 LSt('\<% .3(X >p}n=4exyj;/\G=#P7kQ EQy]9ZĹR[檨's}9~W3d "X(kY)[^?X#0v/҃ UO|5PR/FJj4̿7e%Ԓg8D+率ʯPLnP}55RET[X_!ITRj0*J:-D]񫵼4EX+1ЯcVQ,eP(GݜE 5KwO (._֥/ۡ=wnlh&8 \RurғFL*hm~pL@;Xi&&:I_~0絴.0v]oo&׋uxp'7LycVoο̞o>O#Uo|Yuz`Gc@g%'Bcji¯kl߇+pARX'˙ 6,W!Y>'9)UŰyuG,HZȑHL`8X9E2BGn?lbWa7{;o-njS}QWE|ye,w`:* SD+P%RriYT 3t$`vmajmaB?u>xU׽0R6pc6G;\M 7Oanw1i-Uo{1{3D~8=u԰Tץ_/bi;Fwc*6>AdA~Q!ؠla:)_znakMpkN]6<[]w_]:1|2?9 `%->p8O_g{+|o>\7>&Ve I쾣;vp"X1=F^~u>? ϠI!mEeN)G(##f޲'fc 3Yh2OzzFnĪWTO p#u %i#>9@hb ,P2HR<&2O!Xw:_v՚}[2&g|9=+o.$*. •2h$ =J9$<yI 0{BJkEdk$TINS*hc$&a$1)LdmA!xEYkX9(@!4i bfo ucUDgq`<|DMr;u<>n%eò؜_xjA .Ƀ n51._$ez*dm`'?;Tv&׷۳?c)ΛsLZ_sHu\NVfݢtޜ{v8K-U\^oo3@VVjuËQIݕ+?M"L>cypEFD/_Э YIw>yCay{ZWPΜ ּ*-J=PV`rVLR7hA1XKy5`bzJH:oפ2b}mRTޮ{LK3^h<צ Ebϑ M)WRG<ZB`؊"?j{0T"wyBabCu%jrT5LM&&'53{<&g=>v'6zt|t~; hEgD5 SPf5YۦtGq-%A.cQEBy JQ3 [=KA_Hd!)"H $y2au^P`֊ ׿$&IA=1Yi|X,B#9fױx9Jm,SE:>AzY '1ȁaֳ0QF-=b{?cDZyfƇ&L[w 6Tw uu0]ѕiR)qm1 2YvcLZS c2^vYH4G#cC ]mN·io3KF3t[}bq?Wc'XAwM!)813>6DCD5!wv?w dJw2^Ϩ&yGzG38DN ${n?Ҭ7œen9 3;9SP|ya=N]jq\/0w82*#Zxg+Yf|9A{*\=O<xOb}_RӃ_Ovﳂ)+6WnUgU[ `8׋oUnPpG2'D%1gHN"**^ne Bҭx FX,w^Y7ɁimW<}W"Ljrmu]~v^y'yπ [[dmRo]uOti/W(mTޞtK,ɻ/]*dJRMz OPܹ BX)'wq¤06h\E,7h m(7ځ,9NKQx'S32+5J !fB2hB*Hh4 ]$CiL8&Ur9QJRHY2ba(=JHa?<)j)1ag͜p^y12l8Oh:{7xdy"> &MB+ 5R-,qwWT'nA9IǓ*L-Țk,sBC A>B.7`BJ Mdđ]ӎ/1(ԤiʘE"x4ÄDl0Fj `Nw6gbE&youi82$\2,sGSܝ70wbaY|UxO',Tʛv|&'';HT;,RXBi+޾odsrg"u|LodWBv~|=6}~\sS1cǧ VHbITi3sRhJp6'[7 z&2wxoX­@` J*yM.ajM`SC^br3˲G:-$TbKjjd\vEDN1+GdjMm1Vl+7#Lx/D0sQ!R EB%}g[l/GLhu8'[VcS`5/:}8C}az':&7aBl1} P z6(1,tdfVf6ER1_L6e¢vA+vqwY)zwUcc>p>>Fl9ýa I W7o-~j|jcDZ9XFgWkZxc+*~t=PG_7QMkǢ^^ wS #(nX{azp&8uhDH0#{IДqUҀ!eҀfbcu|窣z Yg[ZgȫNl1j$tqR‘8x'JYLΠZ Hγ< j8isr+ ٻmlW|ؕS;],mvw?l)R3$ilwZ,/)1mK1%J̌Y<|yxxxt"c[o>»c9azv%Gcfi*꜅~c2:_SAj{ҟ7ήj=4Ai$4*bOk&ma[53/]L5zvih˜ 6U|bax -FvuP0uKՋn}?>=l`;,B0x8RŻ]ǻB2)x8w?G"*ꐶ!`LݱX zJtLaJ8z'j"xjig^.OCMK} |H3x $\lRܑKf/bA5?,(41|A=B {%8zjCye[VI"6ȇ҆$݅%G#\$۫39ZV(T3t9xyAγ- ۃc,+1ϳ2"^3s=1e=23ε Z4#^đ} Gˎ ]ؒN&ERմ}Lj/\MwĽcJ$w&8V^(]4Lԁ=oqTub WY@~ucp*go\{x6+}C}Ԝ{g(W\6}Wj! ճ;\:8`Hi!IţJ٫mK|~vp̓j̈>}QBXmk\hbW(m#$%wwGW`}%(n ؆sv]^"B3,&r\ A1<#<0 2~#[1}V?? =/s4|~|:a?fڮֶp ϻssn|vM}C ,`K 57]-PD 4p0dv/ߟ;oof@R-%O]j$?&\>Bqyy81[AɃ3eVV"c!Cκ>%X™2X0N@4*%kr\HWʩd÷o7)moMWߡw_~<*_.W?,>YЧ{ @,8 / xN[+Vf\~x{~ 7;KZ]cbs{ X'rs{ ~+ΡP^UGĿtcK%d'DB婎wRSvbS)6u.hV+}!cKTs9y]4IPpzN@ϨU~P`{+Y(>Zu@C+AqM6 Ðɱ[1?U  / ?ReNSe"Bh :KSLM1 ìK5^,Jʄ4$T4h0`r :^b3A`2>Q_y؞\&#}n$BlxK^~FamԔk!a|]I)}yC|^IA޷g5\3)yF0$ƭT*k@YC E5)ɖ EE.3R(젢 BBafMX` ,E}yb Fľ("$H)u2D6yU3-e*5ژǜ7uo'$gFZ3픍!t?i2Jƽ!@?{gqb/lȧwPܒM8ֽS$=<*`1P9n IDo0^΀č in` pzrX9Ot )(C@ 4Z%DMSmg}yĮ>CW?_b3B4@\)Ivz2 Wӛ[ݸ*\~;h m]}{6~ln$&V7~*{^2YAӮ؍o7v6V4ܻc1ف>=f[9=0nՁ 6̟g1w2^"ruw<%Z3N& ԙo1{5;OhC=3xօI)wSIܭČ 2g@M}HfIߧ컉r=| kS[gW*GOV)o__EomŻ('`"uQ^Yl(Iom~I`eLߙަ h gS=eZ! a/, %:wc sʆ_6F[ɊXN\V~+,kdpQY6='WF-O&WƼ!E K**2|JNU 19 rCT !oŌE0vFE2@H[I,лJYy&^#"fs!tr\F Id9bÝж>:Sղi(S;ÃgEMR) H(g3=L!X8#w;UȜK-Q*4 "WH e-!F`r_U[zɎE9PJ<GbƆ L3XtkR+eDRˤ2h%  CU נLPt !@cjdL3p9 eѴb%'}{P{"IKL!&wK`ʻhƮ&w;$!e֌4(U@NFxI)v3I0A\KnEZslK{1C!;|:tW+q1K!}gX۳˻\-+׊`\w ۶ {<\ژ2 r9'u' //7-HOh)>,7~Zء>jdʠs$Or3_}A@>kYEIsa~nr=%Л$&ߙlq[ܝ}ݿyJ&m zSR `Wn9@٬C6Tt+nq \U(=|cj$Ͼ-W?Ϫead |*jY:"G{RU~hh;voVh| ^$, 1Fόw \q΂Fό By. !A5U7JW:O z KRǃ"wv QQdC߃8HO3sdkz0ĥ[zBG=,!Jszr)UGpFIzeջ-A!q#R< ]gvs@!C`m^VN1P1܍P 7DPtRŃ;ΒtSYo rQm TP\AJeqep7A$wI%+ E!(G Xwmhι~x$mn2I!@0T\=%[ RfĢȍ7@`/$Q&POkͶ}հMWI~H^nAY2tE)DNl_ݭv`<뵥7~dKN_Fք:%/gT+qQ(  :ɷ"J1ʢcZ|QTpEt}kK)8^Jξ￳*q}B|u&II/f2Ⱦ2گcX$0I" };ccЗD}@t9 Vmw $(sʋ'_ρK5QM;OxPAraa3 I)ϋ<~dlLG}^Dy9QhsT&ESg+,XlV8hS-9a}@@q)Qm@~ `)VE%h NŮTt8"e~~HƼ>zylGOC瀁؈0e! m,9W@{'vlc1{X@7ahs;׮AS W5k~@?`#8;v< ta~~+߽zͮVH?>".BRRsvuu `׌q8<@B [h{ DK 7Qfyk ҂YBΩunZ'Mq,ѭ.Q Һ}t8eqիD^x1<{m"(-R]w hQ0ljbBX-Y5A%g nZvIgĻ|Ɂϗ mɘ %wvzrH j+uSPqZh^|-}2,SU/TW_>&~\YskЁs0(o 8*#o'* XfGT xu ZݢT".6y'$%_JIJ.}XRҿr$%ycCpyc<^xgm5hHev3ᩮ>ma^ZcִZK+֗)c0ϥJB_ŗaݴQlB |C0} 7a+KgR_6v@}b5CJ#<ʛ g$$&}| ,-o87㟇*(Ov$/P8RR%XϻEaU%s'mKkRY?}5sr׳'7)]u͞˳5'H j&iΟ$$iHϞ@{l_O' v4rS7KӧK#/|3Bc{XɻL+hH KX&P!<@u+;q|Bwk9&#1fo{4bQ%"4x=L7'.Ä{jQ"];&(&X؛-LvA!-$P $XyF2*wہDd%nyœ !ao(3K!UJ!B* Qw6*,K뜕64!Ra$ *X, 577ۀ3=珷҆ PPٰR8藃昣H67]+Ubd? ]/W^I.;C:ֿάߟX%)6Y']TV1B@HṀ|u!\Z T0XP #-\`;%,5 =LhpSp-T!ڷGӷ27v>f&pQu 9x-)h)I)UË‹ Pw O`F=]#3Nr{qrX)L฽Z E!/R<<鱓g1 C^씖:3*0&-Z% Pe\{Tz΍ XT,'xaaE d cC q:'2d,ٗ.b`C:օE],ލww/:;EKw 3tua?E{O4\KнJ8;knʨ'CqGHּ[WgJ.<`wTf=']4"R^=fW+ |~TKMwݧh#Dқx3q9-'1ȚnxҺsn&~FK]R䳿ra1 %b=UY/sܧEfS,M 1 d`I`QLp{0\@ADSU"=!$;jx (#X,E;sH{%'ݿ_d d`96(8r`5,G}9yg_' jh~f Ͼ;+ d0mڈsa?ww;d:~o|3^"q>띁;ls40q)|!b\8y{$Cbf={"\fND 1˘D,$/8E $) +ֻ)giYfys Lm9Ebfrɀ]nU܌f0.{6]=$zbƣD!ZWFvB1}Ll7l޸r@3\~'=dv2x>{nl1} $k;L~_#g/UyO]8q ɯ@~w!.!o^O˖]ug}dYʹAH0v WʾR J*9Rxy祱t>L,h*\  ct>Z[-UhDlxd>QKG~qp~@NglƹOOtu)*cE;vj1#6i40팙1% JYm}q-eoV᣷M-bo怾-9o;U8\ǚ:[g%IML>fxOKMbAαh7 \4558읦AU w| c . Jn8cTicg(HSFB&1¡T⭒jz)lRӜ~5/$4A@IX:v U&,e xBk:A[]0u5)fyA^p伾ڽE+Ec}G h嫐)!B!iPMe{6"%TX2 w(yXaRA5ЌڱjXj;L>ldW&T!,\p-H`Q]eRJЄ0zMzgQB^MGP0DŬaMЇ(ȉaP* FLV,DGgE83:+jl 5Utw9/, *F!1^̺ү('vb(fbNe+zi>CF"^.A AbKcr"ۯ(r Pu.gB@蒪Lm\,_ϕ1p%6vс]U5<i#ަBa~WŲoݿWszvW߫'{ bf*j!VSOn | ޭ*>]3B[qVBޭrC}[=wэ>=HLZ(@f%Đ=VEx/mYIau6-ZRmNզKiҦ6^'`:8ԕeb2W9܍d{>Nyq/_<l ZFM{b̓Lf磳8mwn&~綉OrE?4i5D8s:L2i%RD=$'6JRit9sr#0-Cu`37 7~J[lp$)(ns`# wm=rܸ~Y`j(J N "s2ߗn{oRץkTQEOtfA@4qy^ٱ t t]òkEUBx`U$i~}(Uhz2=9s/5zf}3k@B+yV%5,SCֲnIK]*y ޗ:#8>mxL;Ѿɟ!8a:SĴ--;گ'FL;޹yݓu!{HHvb.t#]+g1_ãEm/6Gw,sΑ5 p.:Rfӕ:R⨥=Q˃[+:6&2 $afr%&SY Nd0o? IXƔ!7Cɧfχ LnRLvej Z~B}_=kYé=q34i4xIOrsI51erSV, srbJʬ.1z.eL(Mx%̮.q<+#0S4o{.!=ώo{8.A:3UKpGH5zE0tv?(0SsOU>9,}P8VSAᰶOgr&,p%N@$)*CL0bD1LhgM& H42qRͤtت7%\YZau!&% mS@%S%VHc]rM:$|U䥬\%ZB%Fvk|j뼭<<;'˪ZQzAӱ d\  ?鹿cWc>0\\oSۖݶ)o9YW/:ng#3-ɦxAֵ\7qJm^}ӯxޟFPǕ{|& v:@!U yD&>U9a*j# *ay.Ut[xƜE]*߲"VWrXo$g)[DazAwtFZoyu@ ǝ`0`l.ˊ)wUQ~:R9)ۇ'~M̃o"m}]d}qNb{~y<)ΣV~ycyx=P~^SyA0 loعƽo8|oSp:Up"27*o"\T$Ee*̗SnLdo q^0OgLQo|s5SlktˢbiG>i 5nY}joK{%c`j\eN_j񲬞Gʪ *}܁=UUxd|uZWEN`*Xh\j+gQQc~TyFҞJ2F{gRKu>􄒆v6*TuIUc1UJVdUe[k7rnPdh.E䐕h{_yFE,v_@j]FVt2FIeaQ3KNfe\jtw"GXR>_´JjJzsaUĻW>Aæ~U^OuC5vSL"cOߪ, dkxjU+-Җ Օ eZ`A:,jgQ66M%ay:+-=m YlI+64>PPjמ*ITS9}:m嵿CzZDz;=ql"jv&3}6:]kPRz,$zTJJ PMNyU%hkJDY)WmxȘ(жD+|NNXuԂ<"ؙ yEm T{WagJܰ)En؎BYwWJj5ky˭sUV֠=6dR#U-5I83VȘ߼VA.ǣ R5}1Ӳ4fgFKmXB ֹ$=#6+E0ė&y.)/YOwj52!Y5h5cQNŌCC>ă3f*C,pSl͇riӇTn,ISGp6(|9WNB,ĽtXI> U>}>bEQ+\Q5'wciToj?w(!GTN\WgBh#e ^ְt 'k`iB1&.qNry-N_?c[Dso.q6H 7@;ZF֜5Б ;Uحvc s0ĺR$:i }c+ *-qօA5!E 4kՉ'gΧKAc|&xX_Ox @]߰Wyăg76o;C=赙PZ`JW(hCٚ*U6cW`=;P,Ui%bf~Zn&ݣma+v/Jr2̂F{gsgVXWW&LvYָ:PN1*`ӎ AWaJXO7 +"_΍SԶLS%<$OHXdggۥʑ}BzCt1da'bLo)əIAg";sK>8yR7Dw ;N,4f7=/%%,'O=شiL|Qьy ܯIB~C(OWuj o aFI!y_ӿ;'%=|s抾Uu6a&U$ڕP FﭫPCtmi\ k ;z3u #wo?zswxr'J!G=XL[-xUiAbu!g ] ^mF)'ɜakb $ghGU1tB>AhHJ( b2AHx={Уfx96z BTiiv;PGҏyCtZnΨv3)U(94nQ4^eM= l/ڤk~߽~rqqqςUv=삾o*n_Cg{YJRJ~:,)2W˃D<|;ffUP" u3lQ!}PR Enœ B "*wq5qIp03Aɂu:|kG ~$=Z+C8uOrKMI'\|t=) p2i$y(Y0 X6cl,hNҎON#ֵ84nMOz8J^ZdiTsgMLQ%)9b%8ʽ\9 Z!Σ4CZm8%t(7iyI;LFxڸݏ ll6QãE,!NZF=X1'{++]3XϯoX@!xBZs8HtU6D`d*G6^ [DLؗu1҇,]PPFyTĪw\ݴw],ȵUAg?/>< {' r\s} Q=)lZ N'AY6ʒ>=~@m9s@lu%3!MP/s%/}*tjвum\#mIB_fgR}wYǙ 3 &|X/>m!s ߷/CbUUu>zvW"߲P;|^RPDy:/PIo lKß?~?{eO@q9i79^fl\6pxe'ZCʃ[z{ch^VI֬&(%%Gkv9Z;S' }yKoce0Pc~Ԓ?w5 x H > KV|_MA AӀ%zlGv<# Fb,i[-x Q 2) 9!;r`0cZ,}N%$j4ZhmRyK<"1c'4"Azt:yfjN% `fZ<̴Ɠlj*g?~h˼팃goKch"< mdRîrw} N0'}CXJt{88]H9W!_8vZSMbGn]qP:]h"ݎgw!M%gn7Z*/E;'oK7m)H8(ak4nGn?tR|,SHjZܗ>:=w}sv.PC NRBP'&0t$IDj|޿:;h'W|y|v3Eϛ/s?7l<<皣Vz Fjb HGLԒ )C=#R +ɼrjEG w.Kg&(Ѳq%' IW+@' [P3alb t\m#p!3 $IS `A Բ09ω'ZԢO1T; \$N+1{xw`qk7 kO n,mսrKO{.s1IjeG8'ͷNA{%IZ}'*1Wu%|tᤤo7G (6iE@voi\-ݺ@ Ol&E5D[Am,~ d:ZAfCqs&~(D+"Ж5?Tj eQO_ʼhԐKcgƋZ)a U?^0_y"RI>ECSE/p z$-|!/0tjZ 7DR & u:JgOimfomLO[.BΊ ,Q6"#9$K^™˝r1!98QE(@5,B,h~Xg1eϱ'_) Sj2d(ǃ04BWFYlf?E.FL9}g1IpY45 ܰ6&GI6`6xsVbQ2,J2"Tr43^7 a#!!*Y鐐Lk8$d!5pMEHqpHHžB$r}0LiIq!!sB އjh2$;*nzbņIIR=$dZ?~jA6ٷo1sT'埦8wKf(T`C Y=J$s_q݊uvj^H.IՊH"z"lu=g <8("՚҆K~:/Q! 6U} Bj_Yվ35sXEW7D|?Kz.^J!Y2^kF MTaԸ{w3Öf\Z%RJMi(_n)!S,#%hV"cvJ}N'H?7='?M|@@͙_fxy3+\%S9BhQǕfʣRйf}% :f)Δ!enkgb}H4!dLлRȓt7/w,n}T)u&C? q0W#(a/ M/EM ZY~*9v`e 0OϏtrی]N~v>Odz#?uVh+,46@$vAS; ,}z7tߕ8yRⲳ%| 6E!Qoscf2>,qઊGex¦+oNO{;l# ?^vv7w)Zgrs?mqdW l?|8ge-lwfF|Z'ښ?oԞ@ϦSDjc"[Hs!h.яNvڢ'yn'5~NGI FiN:|{sv}wW7  ?;|=*B87K[)Sq3OQ+R*i`ϕN/m4l\1f~V8CԮrޕByIîq|O{;{_[nW$0 G5^_׆!<VF-26JC69aA]&/e/Ƞjxhz|qtRo7C|/J*2˺J^OCU@D& mdt<),4eQU+燊 `ΞnrJׅ." nj5d̪5q57*t^5C3M (1JMbq-D t 6U]L!9b7 e|(|&ǍS$#R̉1`6L\N3QPk3(Z@%ajKbWeԦ4&v-,# '812R,z|Y>ܶ',kNJte6`|olUC聯 pJEV$/d?>"qrnkT>ͯV2xr9sA6jEKG;9r狥ϳ +k UU{:6RŅT>vȁDz#)6#,KnƟ(Ӑ4wOAMȆf%1l\ IGk25 4Co;5áQ{oπmqotyrMKlֲV-ͤ[C-r9yy 0:Z앢dޢ|jϵpMfjUɭ5d7UH0ĩ~|wTo; V<~=xW_t]1y_qZ}*{:;S4avo+uȘ+*{<^u)˞8 gN< 4ɚt6ҭ#XѬLS%':/n=p΢xʜŹ&ݖB-Gn]qP:]h"+HevFq0eIL,}#TjYG4XqU8w[>{ӻ!߸)K6tq;SfSfyvdޟ>s^'@.z?^˛6Tʵ}{ {jqվg3|!".y$)sS]IX6PWI߶oۿ|QpN]Z/a?ږ@ߕwS֓?$wwy!g# (zattUmUHAQU<)'D2ڍ>-tPIxB$ΟJVZYdը[ʪ΋Q[s[ v $׽Rlʀh>>&G5ϑ9XyNU3jtȳ #JƉI2F+@/h٧pH}×OHǾq![Ό 8.mGM5lX5B fC՞)'f+2-iJ5)$9AeC,lmѩuv8j6/Ҩ6 &pF2:G.IkZk5r;.Y5K={bn:M/f`,GllBziJhFFSUvIgʒicrΚKmIuʕGl f#3ߩMc3_S[]:Ư#eEcrU9W ;IYAJR&sX)MR6h=uO'cCas}LO@?gKɮ|Z[ILp㾍2G<<}rU~J#\HEi/ -s2)&0挝wfAGsQTf 'MXۉƹ ; Ld ^t6J J>C7jryv14],OPui|d׋PH}厁awj;ֶ3X86p/pbF;ң#VtZjVy]C "Ǿn]uBj|rYކcWּ]YDsOϗO0j'2J/<<}r> c)S$ʒ$=m \^YKm,??qr}dr7jNLQ39~uKYOcr+:XgWљbkf5T!; U%sU)rmUȾP2F4rGIW mP~>o6YJ١FYprf?pifvxh4Qm -l> Ebd߼iBqg:D0?_5[sUl7FkYv,r~k-,\;I]hH䀇lWpW*Ƕ32|kކ2׭\>ܾ]^]?c`o ˻axCRyHe6<z zRDkPn~_Z*CNt!Y†N.b/ <_Jbľsvx&? t:IO޼&ߜ1˄sƷ)`$:=X7IdH"ՓK~5+?ldèWXxkٝ6]Ms&Yd6 ;$3㶔Ln]]3Q(/9&*t쓭R R-Qlbe_-g \:ATXc-XbLAEJ f &ugـwGO56Ɂ°%Zom@aS o*O"ٺGz7Px>w>ڑhSyN+x.)?)̖O w;; s h p]d5Kn @ߧZAXۯye-ߧvUK_r1@:RZUYJQ ޖ96zY~|5vf$(AL)E2SFL: r]#>DbgRmeN(eCp/@mY98+>kxWkxýna,t]aE-!s5{sv>f~ueK'n>}f;12laشCӴT7_:krrmVӀgfYj%E׸5[¹1uKw^ӂBo33P]@KuI?LNmg@m>-7}9wX^enoIa.i؉U8" 5f%;_bibF#M_,?^w.aPۛގ/O+T_}_gng3`Z:0ӓR,wubd=i0Y-.z B YID TNŽ,$2Iʵ mT.yC^S¦T 3\夳.[}d\TZ5۳Rm`vKW݉ ڿzY{[;K[M~*kwKȬiHcWs[v>_5<Øohf1nc`0x`'s E/1Fͫ2aK_?q#́X5gׂJD VO)J월!/OPlM8هYFJc0)pl"|uвE[ ~0jclBmQ Eqxf@_c-aJ5A <ѡ K/t+L 24rAm7C9nnUi o-ɗ|֜suQuBWþ DŽ}0&h9CEwxٴ3{fZ̒q{="u{q$:Wؙ0O,Ɏcz&5r ^V=o8䨓ҋ&Gpr~}^N|m"ÚF_BGtҖ*B;{ }">8z;3 r}tL,a*bƁB(88[U**5zĬIE{ #(}V *ZoQa SYº!qgCtĺS*եk\JGK%T|:QfB(JrB&jj AQ&RaCRpgBYs"}hgȈ/8F1B*|.c49P%\Qw''rs&àH]qC[i-F&]xjn7OD%T!װ'g"z2FI^ta tZ ȉ#Ff5 BLQ@a+ FAr`Tt),dL)@mNX*j%6Su%5 ):l$ニv5 Uytߖ0faMno_Nj+dp/Ct5*Ce8% 7$I[7ٺDv &.}[=jFdg^֥3.vt]յek[u"4K{JYe=$̀5hG0Խ6u[Xq7#nt& 1㓶9 @.u/i2U]տ"lt^6hϵF4c&R_#lre}5Kgn,,-I?Nbdaϖ6˜f]q KaS'kI` g VT(e`azh,/[>U=fK7KwwuYYr5v/yTE%oIO1Gω/BPS~g22Ԍ:Cox LSa;˟[6Ӣ`"vaJ? /a^Keˌ;ne)<٤5_zKjTbY`VT!%!qc׻K$*;3 ZD5R4p&L &w":uv[CAuUڥP()Q:[ FH켯%[mL(agv?Fބ2 {9i@ݴ(  8A#MߛC zP1;$#M:΂{s;du"LL6d} p9 jeA: ?xONF?ۺ?қYF]B1ԯv#KKP2J.;+@/3˙`Hg9vV'bNTvqZJ8 $8.qЪCE=o8R8Ixxun:'Xuv׽TU6ћ'J3{pLN4;VPifGzw-͜Vo=]z7H;FY:2 CPي!-֗њrt-RM^`ClO@ g-<1tf>*',Vpoa- 3Xp vߔy00\f[0sZalva]~óa )FMju~,JqrXX@toc52u,y}ZkgztF=PXV Q&>~-{>Rq >YBzWݦm| {K-'xNeb/tak0[CBT~y|LEk[4c3`FY_пx·17@[.ÇZEAzgX|cp ,E 癐x[ 7Dc!>v@[zR Gu#6Gn XL9fƒcWNd]lsmP HM^fRFvˋ ь]Ԛۦ~ nO~W5F!ԽE.z72??s?x|~yC~a@.Zs Nmd} EI*E։T>S9!ePߛm} 'g;}|C LǗ}h;GmLzaG[Ҁ.]`*QG3?!*t@)j62EC5e.SEOQНTeO%E$Ɨ9^7!y,70Az2Ę`mGޒ=_~mu?9-*񃯡XGda'*3x/e6MVV(0 ^A JuIWmzp,ě"%t|Cu2jpZl Vk{ {ٙn:nͯ_}sc#J&iQ$?@x ]f7PK\3'7f ϴBꑐYsȂQ̯)driKڴ^Xtg?͑i[Zoj1 ZA`؟z;Gfq1pBG ŖμV<ιϪ:g\/4PNuzRj0Uf(4a,&4b%CEϠt ϻ.9o五(oz\B:f#sZpf`HkU4֌7TQ^D*+ϝxitVƱ鯦H<|j(ndЪ sQMԡ r0_yl%ic`gކkj;Wȫ/cr`s-klj3blϮdn CvJ釔Y-kf~lx7`۔ eseH3Aڸv|)4֕`){E:4|RlkS*Lb餣ZʍVw{ܛjgSo29(T,jBI\=N,0i R61J#->J_(̵~Rh!HZ߇ށ)`,O~8Ke̋ 9Xfʵ crQ >yH%{2 *-)IGK-@jZ,(4OY?U&EC%J$Ԋk{$TZHN\Za\dǛxҗ;-OGU]iDmi]f%*ce",)LaVTBz1"ckwNyH0/PBdPtH[X݇ckA7}2pA9DwIKtF\୑|?]je br.7ܳ.}\Za 3 Mue\yCmٌ"cZT.1r'iNc}uT.z"NZҦ3b4\z>0mökyp)4aօ}y.LӤ}7yIrAPG+!CIbH#9aMާm[сNs!Sj 0qDB$mL.x\jR7ɖ:}b4Goڞ4U?Cj(5ҡiaQ)6 &զcMG۵{ݹo9fCR% A"i%GJ|'l W4rt5[JBp67%{Jwxtj''.iӉcN0]jRX᪷H5azi%y9)slOaٙyEZvu]!f{&7 #6nlb*yVsbT!6*C[hV"̙!Ye s uyKNZUE81qpJkJΎs;LKZ>֪bК-F^,4,Jm4[Y TԐȍOh~!de*Qё$GouF1l^]mGTM:qZBهlRʫ6hh &^P呼kڜ".ZS6pZ} 8"6q`]!/5$FqϜ/?YEr Wg;?exo:k @9N^02Ή8a4'Qlf|'N z1 ݑX0JQ@_R: faƕ''P 3uFkGBTR=9uxx\[b!!HAQۖ{he! )WzCnJ̨ڷaS(:I b 1q̾C6螔8RnFyfmD>4T(gsRBIӵgH9P2*WSMXT1 K3R19kpV|&moY+ga^!;MCFhކpnQJi%0yKg˩աF:oZnϕ9I롹tR}(~j< wU, 8V6^uZJ/cd/Z!3b5;2o[=RdCѼh~lf)hMљJط"'9?'Gr˨Bfva:pQ;i[S ev _OXl(Lvk\&u2ŋE R#vףKzgҤwosA=@ډ7 l-4mľqUuA$7aZw9-3 '8YhTɛh/x1J 6}"(VZ8p˭IQ%iB58Ǣzt˾@r?h@Rmbmsm/G6NM=)2Qj菙o~ e0~>I{:4xd̽yzA1JΖ{\ t(Y;̉x/r}Qp #:&JL@)6Rk6×>N{s슸_?[499pIpmxO/^1}gA~Ƚsr-8Q"*Y)~̘VK+@wWz/ɁQWga/w,V/YaZ86PRõ)Rfh ÷u;-3'뗨,Ff[>+e8jDtwO_Dx&ā5Ѽc}>V3PJC[ٙ[NF[J + ZM'LEB\zP$/2X_ģTnٙx&=tzÙ.[k,smhDMs3ݓlVE;stj͚*2UB3/"eMmo,p/|=J+e.!H.G> V.bhxs9;E Wٗb=I_bKiC=hCj[|5.&Hzu3cA;1|nI;VkY"A v)QL8@hT܁ض=h!WH* =W[pǮaQ"3n[bp A‹AQ~EŅ WF: 0Si:‹>҈4= -uHJ‹ro'_Ӡgk>eEJOz)Ț'‹2o`;9dZT mMev J.l8g8/wgmevvѫ\0=YkjF.8>oBwso~IOg4RU&B@ wJZ4ݢTVu4\(1'Y?65 ~K㮿|r6 wo.~5ANDtmtM@4W!mU>a7|xG%|I" 4"KT NsV'M"!b{=[=>h Pױ>9;λQxD_bXAd"19A!0-xCoZm>^,ΣQ*#S` ;gj{ȑ_moV2ߋ ;8 $3Ͷ5'ˆ$']bK[rKb_NbX$*y\&HyHpҞـS'$DHPɌJSMMpjCH昧TKRn"TV0+Phr2(x@r > z)γH LyWGPp6X䫠LJaLݷϋ ٠BaI({WO墹9vr>+sB3/Z&|oc+ۋA6+>[hxL0-c^~B\+;S  ղ8]poAF $!XAͱ{L5 DDEL:MUG GV 0v uyzbYAҖ眇KY\0oSE@4fƧV f+ۮTP|Wpu4yd}on^xC"Mp4n <ٻrd "*~|ӺB<4.{oNrc40yO)Aek??fTF2*G5P:bxpp^~nsi;RDEl$a-ih.o`~ʴVw׿\uk5l^6%F׿-!R` !څc4 "꫺CĴEqf;?w|et룿CףTg2J3M9aRc}P )h{$INŸUK-(+E&U8pl%iTY9/3$A:\'4K ͜"3ՐZRK]IjNOМRe^drY𩠐zDzFlHI;DKB ?^jqR2|=b$pH@yF?ZO=Ֆ$kİ#58t 3 C}0F@!uHJŒ. 8 dKji8$JF Q~QEӝP ;9*@7hzڬLCcp8屸Ț8u;s sdKnϼK4}E )-=%=ӷߴWn+CH/W5d㪿P{hoۇ ܨNZu J7Hz|\0J:<=7?t6tc.:lh*l(d6plR†*1Ҷ;'t]0Bj':w{pwR'ӺGΩD>Q%ܽ?޽tO0^⽮̝ Us՞-*NwOĕZ~~$ b {>N-I+MCQEz4p9淝#T3N+0xzaJyYɢ:5N݆atoz `GHq^ ]bנ%^Ѕ2N앵sԽ}}tmW uN.蜹Ww̓D ׍9ϳjW+F^թl`2&y‹'mbV>H% YW̅BQR©qFOca-0S?8|w0f]PbfX1FwLDZsn3WKU("D #B$cgUDCK!Kg۵cD A &0CFI-'1%?VY$"G/_G_xnۛ;f`&8HJxwr;SI9TRk 9UKc郡\vmfy{J _^r1':J嵏 L>,Ό˘ ;/k"7? Ŗ6m uĝWũz|]/ZjA;OZM &fPdE~Z qDh-Zfd:ﲎ;,ak.(Q&TW-؎Kt|_pBμ4"˔RC)T!f=b<'pKns8Lo3Y!}2/r* 0:/b#Iqdkv Q'x5DcFhbc%lqOjx32Ė!w`zyO3&9sݭ'Kٛ/cfqm̎'3?\<<..~ \.o./) F_z;YśA:= 6k[Nq9nty"g>vUd]Xu?ûw/+//'^T>jUP/X: I0I, E@`i+a?dJ\tA&q!8 )ni$_{Q2|H8b9P܈mާ8~Pݛ^:65p%EgLO# \I9WRY.'-=]EY|}9NGY>J]rz8RF+VdT*_AqdqmM5:9\ΖevGl2+ bGmEev\2(oFwj㧰erct L2B4eM@YsUY%5KelDȪ{KZE@Kvh%{X D(A#nO+h )Dq7TŒrnҧ9V'R(hazE7I{pC#WjŔ,T?.PoL=}}v"k HoTx|Œ2VCbMN%{=i>J{!Ew3Vԝ`]I;ҔkJKLW FR0I Psi^nWx:K<^I '1C8Ō$C=M`u#Lz߫bQGlo(nK~H&SZ=KnJ~l5~>lXCMskQkI6\6ֈeb3賙汽f7@Clc;Yt2.[}X}w&s&ztP*)KbL4QY5g)# o!bS+RJI(+bvJAlf|Y8I`vAF.)Y1gpFBi1Yqɮ e`E{0b' *[֨fk֒dHY1:b%5̊meVoHT܈mGyOpV=~6dFJbLzBWUzQ i~³OOY;TT;rE5F7$@zn W_ jSRojg(Һ HlHKv?!G*P lhaJW`rp;|,x"b^8ttZ~(\}|gn0Qӯsb$1\|j:WL7.1:vXLp]aoQ!ө4DL%*SdRe2M%c<@7.đ@P^(qLKj $LFx魕.$qK9ˬW$%p2YGE@W-sUB._UEPjik rPNqx+t*+llW`:WRjB8(i!hRUU#QNq@7T?>`pR(B`LS{(PdY7c2ĐRm'J73YWGC1c7˥}<ѹORƙUШcsi/h$v5F tTIW 0{q߀TF޾ڽӹgKǺ+X2OF{l˃m|eʧ&#)R4 O$HK%V&H%׈]p]ncq00@|O&apip/l-<̺5΃' 1_Nj[\#Ж5'Vb_cy߯!GsaԄ6y"ƒL{Vƥp^M@1˜ٻ6r,WS)9`/`a ";8!+ /)qIV,$tc:7$ـOBuW^aR[+KE},U1cV3C \yˌbZ괡^"0q|aTY|x+ү@OnDs2sus{柷<*E#o)D2<0Asɐ`Q`\J c ۋ=\*>Z) Mp\dm>Fk%GT%QsfqKKtqtRG1 8B'Н#&#VO#.0ks#Vd K)[UCs_ oX'+yLZvj\3v985a{9?cӯ\͒9Fם9xх|\yqcWN]?>1&_=RfNb!I/q-F # 0*+Ų:PRruxYQȋz ' Ռ^kbu J X]sD'5gP( `tKWEM-l>Gm%aY0Gg%ڒ`4b`Yq@ ћ5޿{`b}.Ј(Lrwmoc{zMA{:Xzazk)W9ٕ^D0w-Ǚ_ ]>ƾE;qp"`:I`ZJQ'U([ hX^[ *inNRsoyC#cTw7(7Rz1p9Jft M=]%;˵>~["ڌ-y'TRfAMpQyNIB+*34ֺ:_'Ta>OSiT\{ްծ7gȁ#MƁ?_כ3PϢ3.äz.$+ӾYXcY;nx "HԡтT 1Q[cEYxR[g2Ejf\&D*-;9I EoPKxui܂Nȗ<Zq}H5p.l8>K3Ih%/C>(\A~ Z<]>XzLrA)=<=#ƣ]P.(R,jPx/P,ŻCA|=ebΈoK@዁/LBlHe!h] '?y2QH̀Fn'L1NL?-c =< -Gh%K^Wh@2PzL#=w"o K~|bÝM{~|s=?9~!!oHH'n2BBFN3oD7rN5rbLrrO#[gվ||v8aēG3h jVJ #qcĩ|z\IOJ}zpO{M*TKZ^T(*$ʃQ{xh ?k$P{C`ʽFG>mSi@гk\F䶡` @[1׎SNx煌_M p|#K/5u*HO0JWBlaL$gȜQuL;?xk 8n_/Էcoq`Ժ>Dnz~5~;նj|sj[/Xg`:lӓq7Ҩ&!S[b08ϤQ-eKOI`$pFl ^Dvc'M??Oܕ0'I==`f8<_O}L'dBd2!f,ڙt8"5R5S`Oh] .ҫ;k0@F@Z䍒Z(t>4įB{ efI.+5\݄}o-n+w?hba5ȕ}gaU)JďU_y?\!;FT~A@wv(. 3l볎6XnҍT̉Cl W58–If4-&sJFl7yFyTf"4Ɓ□g4)LPt23fxnezeIc>-,yd6m/EpuTJ%xk,5{A nJ u[*m;L! Qޘ|a>+d50v["ׂ$.kƫY<׸g=5^1HָF2C%J'],}҃ckO>gK/!QS1ǼLVČI] {^Vkjу`Pp)= V'D< γ!B`UChAcW[NXt^'Lt^z.y@.JIЎsl7N\`cCJfoSfL 3#]..}Vr_c_1FyHCSc#k7ܸkxjS7qK C6|HDM%=}&淹/ ~^,>!Za?s˜l]l>G=a/ݧ eUo>Yңy*c_K`c6cN=Ŭ {Q3jsNca<LNT9qva}u# u^#{A!iLYy_&ƈ{*\E&% b zk'<aΛq/HF|HAJH~yp=ځv-*u,טʶ&@xOc˼ry "Fp,N\-J@\.K9ݧVJşJIj[3֍z]I| m[4M+&*ȗ^K'F?<.j6Jzk <:QNVxc3;ÌUL8kɕfJjMldͷK{QD:sWlO94GL2y.h"OZ(cQsai4 @d)\|:2++5HKf^UĜdi *Ԗs+(R2,BPn'%3oAo,ٽS,Kd([$ç{KZ,#b+Tvhek2FۍoU_mx҇Vؔ7Nf8Ss"AD _F"Nx9^y @CHV mgT3S 'dE2^uG$:$xI!^df1f=kn ,k!1 '0^-}ōh* 5#%Z:njr^bfR:l;K5;'F~u#?ܙw*DIfziF3a-eVw }< 㷷D WM|@пArzC#kl ,<e&h8h{PeReReRe[מ#k ,l5gkr^ƛI䏿.}QQ-0f?O<'≹7hԫ͠W z"r`Ժ|j4Xƅ7:l1JeN@cxphkIЫUHtH(wG8fڝ-'eG@f01#T\A2LSg({=9 ]*,z $d< 4qS%kTi1ШxMT mnGE/{{\\⽺fK\$&ȖJ3ۺ~ J7 HLyb"4F7!2Rԩ6 8Tbi"'AxAIx u.4G,/ F h -YQlTcŽ\m+8nXЖ3bۈ71 `!|qzz=owY-22nUx \Fj9T!nW%XrL̖|Ez?FD$&K@go`-ղq<Og\Sq7ܿ! A9Wq:_iy*I%dž:Jd<><*b-3a?aՕ_w #aOҕ+mqpL4{Fvnˣ$|Ji"D4q9TJ ߊ_ vOxJ/"$ńNqL.6JR~M.2)&:Up@(#%iF Y!SXW{!?`/q)R,TCEEJrSa9E5 LX[PeapuedpŖq2d<`̩#_v0nߠ9r,{S5 M!l&J#) 2VgK#1E9U|) [Tzt gdk ߯"@o\2V]5P&x$NZ2o]:i/ ];98n-D>nM:^0]utp{9џy!.RHthR₧Iƹaygvws ` /y 7[<:g~u:EVX{2f}d3Hv*h6~Ϊͷ|o~ӌOCK֛yf˧;G04kB#%*Y X;(u.ݯH0G*ؠ)2vy˙3ýns [ЏW#i4Cj[eN\֘r+;0[>ed7C{z~iBY=Dq>'ї;jS-@}ku|9bt_no=WVjIElTSĸ|H3WexGܧ7ëyH&x1VGRO{N;ۧfyEy]d'2L98c4ܝӰzAK݄n󉝗?n荁s'ray]9s튔x^L '@\ I+(RƮv+щ#D}PҙvݚW.Q2%I]}&4v䁗Dη HܝvrW.;2=el<`^9ئNzl}ïik:.&up{wsbzV!Nɯ$XղBe>{Cs!IMLC(4*"!LڿD~4ËoFrvjTqnS!k]rgVyC^w\rkӝxreZqճm:}pƄJ"x牌u=Vt㉌1kA>Wem ж >vY٥ٚ$,8&" 5lP=HM1; NP=&GWO5sUakJ:>SaQ C5~SqcMh5^B^ȔoGIxU[,|D'vv>g0nMH+(bv$V ~QE#c5. *LV,*ΒC닼wQbkLJɘ$iy2zy, 1NY-y ࡒnn mSrSk42DHMJKo.LE.2C(;> yJ_BFh*t͜3LQ j $J*RZ]AJF5Q$ENE.Ћ9BmK5υSیi Pg *1pDRs ;g<)VUHS#"BnUPfDQa!S:;͎h8/\&d"–@qMm~vf%@}j>D}JC2fL6"pC N]d/zr(yպDe:[p{?aE}XLo1CZ]/;:Gx07O3[<Ν/7I6ɿa{z(_3@ ˞f#(Ė' F*EL*eq-^7[&41X(R+quOkp"f3CӇZֈa21byᇺvόK&*rFyv$ϑ0` \)("DʘL8ل?78VzVlK`,!spUrUf;MSvj3}4FO+ n, dk eg_k[^st ښ{wo7E>xv1HѧřpʕsCoJ`@3T(P*;@RIAoF.GARiT jZ}'0X{@BI#f$ J#JWðaiQ# ,x   A3ϐDP_MNMB ~2VErv ( 3Z)s=cCjn6Fd!Eˆ\3 DLbuo@֏tB;\ŁdOV鵉Ife9$R ,0L'ћ@ǞtK,iO;3ŧ$&d T:.Tpᇟp2g 8Efg⛋\&ceyQE6t_CҤ &f2%JCuJHyrduըSܬTյ )che F0\ʓDyǻ ŇBDl~*`I!RQtvcO0FZ[+ʓ)<9~<ńZen3kIIA9ϱ5V! 4l$5hAvzEΟvj!+<:RLo2ךݖsd ? [ 5yY0DɐHa`dV fߧNf箶cLdC;6 ճSޤ;+{ w#X[yD"˧WAT<þE^d4R5Qy5TNKH9H4db'z q:Zx\WZasz0*]7.|Z"&=6S)`78` j$Xz,H\%y^1ir*qTۡo8x44 +FlKܣXyJE'fM*DhXZN$--Rak¤"0gT)Ӛ*  Lk .A(H{ Q6;)~qhHl>Or;OtoF>3ew,gw=(m5Y]V1[|g2/Otj+|GVw>oqSGnMfߚdo}Du酗eĤI+ylېCmvn*Cx}f5矓/!#Ӥ?;LynR3Zh~*n3<<]rt4Z:0^_7y—Owa2 (,'_(':m3t+OZ#RTOM>Q9xUg<8>!}U2"S+qIx.u$i h~h7& <#BB!0lD!! Sc VRE a3!%!LfT*3)j5k!zvQ¢wq$0+*_|uԋheIwI5GV ]^Jsx>&·.@ұz==ZIjVBT3=A]:;b2m[hZN Y%ЩIr7D;c>.lwNU t !$cü T>gśˈ #c 0%pƟy9QUwj)_6bwm*g_4>8>畸}+qd%tJkAU>JAT9Ֆ|-0ENe;xu!Ge6t eW:6!l[Oe(tB/f![zh< jT%A]@p{˂k -9jK(Avٵ T .mNI"Uy5'˹A[qa.'qjӶ-e7$t[`mCw9h:.1m5PP}>UlJͽguB2Sk'_Ӆ"[?W Pq#cM2;_\M߯흓;'wU]F:U0TF,2Ga-!)F\c<7eWbo{t3p%9h ܿ[cmeв (`knX`oi>]cW_x )ZOi3δ̌#% xҊ"" H4ˑ%gsXB+Kls"yj*ٻqdW #,d_HY4[9nWvwe/uO+$Ub,XaCHDQυ&sİȆX -/DQKR^HrOX},Em?DZzIّ q6n뷖2cuXK/I=;QRX>g\ԳcΥZV>gm)'ZzIٱԜAKܴjPKm-R⦥ԬW+jSK)qR+5GXAK{Li)>F섖2ᦥB Z;-M0SS\Ww6{2qIݛe?(y2ӣ1;.G7mUC#Żf0{Ӵ3L!:m]}~o~v ?V??~g4s7['hשwܽ= &Xzf5}Y2ۓf103 QDG[?=.-۱SDRىt%Qև(s 嫍Cn=Տ/X'nbmb3 ݢJjL_wBM6^t| Z粀w^FWFPy?c3'A)%VӘaSf^ ]43 Aby{sf7x >hWsާKlݑQLM$&P.C%Q Y @50.P#X/CX(8fm!+~(eXAJP"ҁ4پIɑٓ)/0xOhTG&oC!dOyR v2d8w0dVagafr4I< )fYavL%1;0Ή ~0=٨$ qT?2e~^A:]jf? }p8ezSEwa@+7cDj:o{WZQlVC.@W1JdF“#<DʇMdӴnw4ndĘ>.̗gN2QگOx4OΨkI۱FCa9uM#7_͏861gq*߀wj@wF7ل3/IG˓ҽ+9>ݛNݻBe ΁6+Ƞz\8 : [QL4^@Њcv<}t1OO-Sot|EjԴ+Rώ8W]b)cL 8:GF@vBK/I=;3<\K=!@&XaMKv,`JP1F*b,&,#|eCH G! A#n|m-BV*jآZJD2D5VGֲ:ZA9w:#k,Hjv)VrִC,צ`>kSܒ!ϝTWpmv)ig0*˞I)oA՛OI}Y)=Q/'D$B֬f4"d V]Ğ@} .$7BNBxӓjYbnW]Xm,n k,P&FE@EU`ss7N9A.S) yJ^ϲzƙ$Lb tݝ,'gny0]=鸺b[OӒ}=-L욍}, Lѳz$%"wxT% [. \ 0kEPj˟X+:VX2 /N! "Zܰ$]ZKג]=vO] zJIU hHEIŝ!7NTL~a|]ͳk3~ 7 ]bf3r^7MSֶҴ(qR#JMՊjGpZQR.-j,R6|Lt#jGՑtJu\]/|rTy>$?$OB|?xpI EΟrŤ)F7za1)?9W0yw"洶"GY?{sè̻|¬/K iR^rcT|ˌY\H)z$%lf=ӫGG 4hn`"@H\_ wqW =PWh v,W vfUn:H\Yܞ^"JgddIi[wZ Vكhx0.J޶W$ ݗY|KJߥUTqfF۽hTB^ jÛ#e# xGf=d4'$"q]fe A+4($N٫o2jMLr"B+iߢVNlZ3VfWnVKQ'| F `ژTS#1gڭJ[JihL6TrBj㰭9uҪnou5O'G3JXr8յ^%!:A[VCаMlNLavkDN:9l$jYzs7NӊipG)r[0r>Ğ9IA"e?=>qs3o![{`?kSmDRr*)D#"4W}?DzTLpߠ #a+$yCM sI(ʣJD}I|5XVAiTT"W]gvYP>N2NT?n`x滉QE zCYAi%5z,Np$R Vry/x9ГIl)•? \P̃<u3\&/"<@z~ReCP7O;r1P#&(+U]8#JC'7q# Z:DFm} '(w.:rzt@QQHGZsK !({iJ t,x<͈ %j:G0Lݎ30z/ӥ鶤sqJ4v^3(Lw{Q*1LaLՏL8|t:k JC[qʺ7Np*=F"-*=]m܌QGD/SXx3Ǯ RL+71S'ww:v۸ 9KUzFV2r3JFNdtZK5v\+cTQ%)UU mtR.ͩ{gʷɅ@dXZ)tJuTShʬRN~cQŽos;[rpi%pXLH0'w/ZwH/RCQ8ϕ0zģ=P̢˜,WOt2]SԥGi2ݼeԂj ޲>zq %ci)Nr!1FTYPeAih~XIM=FcԊUz&oRn9)0!B+*i fV,MG|m+0 HRI4p" 1\2~" VT.?QQ0a?U%| R(L1[tu+{ϵP$Z.e*qJ]u A6E\)UdJ]dQV JE9mu!xc3-d7"䒷-y k?%]Oɲdn)îdfjf>ŗ'48ޖ`͝%tk߳$_}qJ43{D6l9Q d ҩ}_sK4N%ag*٩y#:ƙiVO&jrqL.nlTS y()Vu"@(@\ҵq8R^5/**n![+8Lf]sċ8e#0ʝ,զ^.ҽ+9QOBKs M(K0tFQ5Fmʱ2T ߌ%!z V#"'=AOa҆&g ߤ~! [lvg64ڌU%X0Gɱ,ޯƬ.֛aPefI2(ꢬ0-6G&z!}/'STh*, =HfEHhEx}< Y6U3*|)zN#3^ž;ܜ8QðQ_ekdO+/!! 5ECk=C`22'EW'!WT z J6*'}=DgMųA=lo2mrLD7 b^nQzF3ԯsZ惝)DT?Eb0ZbPWN*ă U; qU]SO*0^qAaT/8FiyIߞczf]H/oFͻfG|a>Ț 74@i*hŜ nsXX^E\Ꝋ4|BH݊=.` 7U'\t_)KJZ î #bĹLfC U^<%Qt}rIN[!D"O uD5>j  eV0}JX>WG\GsB=Oj@Ͻ( $.^4咇YNdSJX  m\+F~wo_r@>,^o[8\ERoصn %ٖ%%I.%Z 33'H&lD8[*>WBÝ͹!^+|@R,PVSp|b4IW&2NH7Jv5K78V+qgU#uԚol\m>h71Ytϣlv  dٱ&I0۠PrFQۙ0F'"I1DqzpYP14T=tRH>H+ӛE2uXZy*̂4e2OT0l5%s98a#"dC?LUZiU=bLd=@1}芖g8´7֟zwf9 Pbpvgbtau ygKӨƱ 2Z) POZ3[;`W9:moaisĊ˺ƽk9ySޣmʾ٪{A&Тl %u峷,# 峵 %%vSbiy %FZ Q>; g>A1hi]O * e﫛+sA\-Ae'c>&!yjբ,|]-\q8 c.owRQ NsSqdr&0 &W4\.|z\׾Y3!}y1L^ i#1Ofp^:__O o&_ *_T?缬${y=O0yiԞ^O$d46ܿ gww^C\23ޚXy }_ ͳooz5>M`wx& /RB^[MZP7ܬibG7.}2W7׎B*zecxAj^I^kYN98pP͉ɣ:8})^"eZ5p޺,8h\+-̉e^v;"6Lp('> /~ۯ~r5_ x?o+fU S3:LM\С+<5c f߾0ˆ9/h;wnx ])fc$P|?&ͻ i#h\iDGX|$ ͸:`*nmzφ_rRHlՍ^I[{?.=H|GG`\ ~RȽ ~AN`yrs֣RP(g'cg<ޯ'EHJ' ק`lZUa6=\/_wAl# -̥&Hrda tp'XSI8qSp\kxVئ \*]{HTfj߽~挻JP 2HyK(m%Qah+W4 i9[]A8 v;vH&ss5XXi9[Cwf[''pf;fz~O?t:a"{}zrΛۓ~ 0qb^fMz,J0QvtބɂWƑZ} EdD+;bo!b J],ۦAɊr['9A\ Ѕ^6fBI~데9ȋA#/F^T5| F]SaG$($ĉBHp0Sɑ+4{MG_e]tp:M_ɗ; 1k74ggW^5\LZZort7֬l[Hgy2&{BPe DBhB7(F Xl?[RE(u1+6b GkMG#pcR N"EqPN0kG b'4R hF ^!ğ9cٲbj@чnZ!@܃)Nq*fM!clB1FİHv. Y OsI4ZD&5riQ^Q{r bǾF  0b Z{3LRk$u#g驪bSˉ95 -kvUGkt`2;B HE~+eumA6|r=1Yd&99'}E຃8R0h\VI\j/ 9A |]>""5=8S5(2WV8L&|߁ vt˫Aw?;k2|w򜬰\F\0}vWnP<V/7bf{cw#kܟ6i;Ycw]s+pۢE 7M`mQ-^}@ۼwՓAn!f2Zܫ%Q&1EuȽNu j3947&p4UBLp[GjUoJ|peo -Po5:mc]}ۨ- 3 ,3h,FT+Wj$'"VjwZ*-vqT+*^џR$R[!-(cRM67CW,Ō1韟/,\cw#`xVFb_Wqxܧ/i&_X%Ϋ0?~ZeӒ|*DT%$kMukʃi:xAiӾ[Hֆ|*Du֔1u1-;uk~ Һ!߹S2uܴrAjHbu11nJZmUb &u7*#V^$V^=&8Xy{㑸-ԕGVZLVG?+e.#{8F̩j%V7B$G( =J8Hl'Y;$2a)AT$5 -66XA9%sr˾W#n6S?\2@a7lg L?SX 썕c)]yāہɵ`c>l˩!L Aac8ygy Ge|巄K&J#VJL a3e|XRC5J0,Qa)kTSj.:‘`.Ħ>c[IiJAp JS/Q$n;eUC(kE3l9*dԭj ʆºHD IY'r PqҤbQcyl̲tEW@я hȦK"L"' k-H?9\)dڈh.DxәQ@e7| WnGY&Ѵ.K hh} !!O&Z( uG4&.LJ(Cwqr2u;m(!9n2PH=>f'0MED&M಺kakt^t⁖J٪ ?uE gQL{+dH2q^ U{KF|]^%:1DLX& FT:c$bn$T«Tb N6*#ဆ!>k<&R:1ȅ7V FCIC6xR(jyjJ5kG{_,g7|_ԯXPܪYpZo ceMw$k=*1fBW M!ToxI'u@wCaA/ꦗDKVݾn|aGe$H|#Ұ(lj=6u;DG bHTb?s L30s 4J#A+%31"8)ZcȧT|\꾟NƫuWi m Ԍ7pҫaPp(͎Pn]w}[5*qN.K<ꄛ.hi˜tkܪ1ئ)?:>hЖw|"1JLw՚} sg@<lOyT>~EBz_!7AΑ{x&=%V] Ǔ8$"FV8!RLpR J(ש3Jr*)X)ߞ_Q}]b{]Qk136b!5}4c0 ,' H` /I%IEESTx,˵0FcFR*SBJ18A>5a MSMuTuX7+-v{bY9; dw'Z˶GșҐyl4-vX,U ?{ s݊o۸<+l+l؇P XhTZxD`*CVnMz\wNؑs5BNҫ׹Ga!m=C,$1EPrkdR@l$`MNzEFtÍx4FDx D$a-os! 6턍_si]d-!L!y/:=BXepsT&G&vU{۹z"[JoNpЃls"0Wan%E5?v^c3mvսSMZByF7;*!|EzTk8M[R u_4tQni?Ii8|&˦>uzûQx2H1n I+&nh܏zx~ӻiGAYht[4m Mtզl vM&ā!N_ Ja+d'֥񲧦l&62D$ByRN(wzjfT-E:IoS˝2>?59jGmBxH䳰E%\%/5V !T2$ J0 @{N* J&rOuRn~oAI">tt[k, 5&sxietA  4^kІd{;R-'Z+-L.A#֩N2f j<|g hg?7̢ 6Rԃ%[ZI "p~4Cbu3-]4u8q5RU؄X 5YIU@?1'rm2z׫MpqTGgFOx$^b+faoX,Bݪ.X6Ϸfm>oPŻ1~qr݇z'+r>~.\_ߖw?Ei _PݣE/{-uk,bƷ}Yha03gɼlou'Ec$I;cal"47;jR4& %++`TRiC 2$l늛^SAtWT[x9llQLu5ϒVzV*EJԶn iRYjmUqUΩfѺ>[)ZJjK*ї/xӵRi)f}@5WMu< _~nW _Wpn06]~}[x-XwX(5rz}9у1c8tىj -[Fuy\<}>D W 1]cBYcyT@waKyf{Zlb-,UƛK4($<6TSUW 0!9(tz˛D5_ª:PTUSkc?@򇔷l<{HdfBh%(ҙDx~`<ϦӮmO6q\K`.kވlG6fwW~p}Id${'8޷'r{J(m+EXho$E6<\;Ե-e#)G~9ǛEQllZQEo.MbHZ 8#cLz͒X azʵz-պij*x{L]5rU[1H[uEv$&W9xŲYR44bmTUjX(dXr`H%[&SA=9U9 )r;7((cʦ uD=T۠8'viЊ{UU;dlX~=j4& g [k%=`6r4ЩOjW2&\Tie8>BsH}u'SK۴U t.׿pLɰ32ck'ߩէ&W_RɁuP3:tw 3xzr5n'c<bfzҥ)f:[8f_H BQ%aLoIL$& fXgY]ƻ7wӯv& rtZ֒ $ ,a[ T(TUʺrJ]D(_8<_Ź(Q4] zg*΅_ev@ژdB%2fBUb`T|ces Ei.lĊ˨)#lCc>z0#h+%bă#HJˇO,rVn]}r ''=O^1>9C 3>yɒ!Jd|Op"q7?Ŝ#^Nu"+z"L֎ mf,TbvLD9c]iv;)) U S;OH)0}dm" TZӜjn<9bL<83HFuR㝯vMÄŬRƹWT]V™ J'DU0ߜiZ1LЭtV3md/Ӧ4!C@')MDS: + VPjc$z|H!-"rTԄCqS_5ZZ2~p6LSLtAUcTc" PHL`~NhQTm "ծ)n0hteŎw)NgBL K];RRVYK3A5BU5،)lWx/.DNvsBd⢷4!C~p28g;{F}2l4ӽF* @jɯ6jMhvViW)f]lJ_Նx~*'d®Z cχeؤߖ2q~_˧/֣`oqfwbRM!c,,xR E;"___F}Ӷ߉֨"߾(qq'uyw=^pJ>D+T?>} cEƶ5 H 6Ɠ3$H<&}Хx$_3K1u,9YVqD֕ 9,֤Ȱ|{Q$یLϐf@(`now/Y"A ܌/P@kv`H-=ˆ;SeF콑eVB(C6Ę17IlPT$Wxs2nZv7g$)1x IS]^C4I=@92;S- ໄuˮS@H*>|n+qjf/ʋU_\u?y4:moݮ{0{U|~We~;.Y3I8xERd '!QiZp 1xM&'LK/c!ZlwcPy"V\=zWA%jzZDMOzxj_BJx?Rz o5+CDb-t '7d CV;Z[6w]{nr-0[JJFsQQ VbҀ%S81t?TPLl">Mg'D ɔ!܉2eh;V}w34ov^w O3dk26(lAR͏Loa//qZ]\hCηuHq>6,No5]6ED7ޓq#W-(^=y6 n6۞\+i؋KZ3Tb}V6Ό(.مgA-iM%Cᥫn vatM]xU¼ڀZx=_ڍoU%Izծ)nq6t_N~S+7sm7 |yDM\\֝gz;.sK7y~o埬;_<nVB-Ӕq\)~L)-xCJ,g'P!!߹FQ879n45h!o*M6 nmHw.5 L:Efgd0O!3hQ=G M)"dGk' uM0JE*Z*aaQD@s` @J)x}to/P_ހ1Q2Fz .ZXͳ%"HD[p%/<̱<|7TU4a›q 8"^iׄT"7r7/G.h) G{uC E7Ŏ›Y`-FHDѼZ U#`ʇ>` SIK›@5]g! hr1҆40=F>jExl1Dު])-j{5WC?7[MUxs2"8'Z*U\ՠV{8AC;v&1Ý^}c.m6$;(x'vn45h/Q_7P!!߹Fɔ54VsC.cn45hoʵM nmHw.5 ee,`rlxِKܸ)L!>P_ա s >b¤sozl:bR:ԮYJOZJ Rj3.(0)%z\4,̍fBJÒ+Ԥ<ljOj!kiRzZRJ.QP7LH#P|/}Į;B>Kb3LaC+?xigYqYK+*/kD-'EP_ZYG8x?ܚR_|wf1Iwe@(la m䨞O0kEG1HxC픖"{]H(-H\LixơX+,alH K8%mҋ%-zT,/>W(-\yfJEl".L2\@@x$8`7X*n|&\܂z.G;3zfL\LR4]{U]|Lٟ\D%spTDS=9$8[D"XIaWx3?nZHv]_s~2YI  TFDh.άT3b mHv>DM/;d2oD)x#XneMKyt`'^fix<,3od?M;Dae'~Wo_L13w1r'ɒը_ۋuzui"-g50q ZH麞u3V3yy9Ӏ2,rQ(kL` Is0AAQ!pQLjbH`7qɿ:"m #PL @ΩjuRܵATBiJ۫EK.:*!zMU?3x5j'-ѓqyMY0f5KlmHo+rzE ljfʏ~u; ZdŤK>ֶ~iĔۓۗ?X)HRiܤOU\E2F)#v>%y39 )ɖ$'ybGПބOZw(SC 7DT; Ђw9zДi֕.r].WѼ~av{s3xwLO:SU0XU( Zgj F+uqTr{z^7K<iH6%[8qZ~4""pD~s $sbB+0;*.Wd?V>oYQH) ~)rRh"tUPRd0u&DNir'*3pJӴptB( ZZBA j52U-}P_աi-*9)/CJ>6l@W!UFVPsRzR!LJqyQCJ)0)-朝K&DR:?Wu0bHzCJiX Tx~RV=SҰkNгB l ,c0D Zۣ-66[&Kc~Zl/op>lxJGP@#Ȗk $d +>G9hc$JR=c5i)Mvq0g^8`rdhu'p `J1Q4P#Fnμ972eH)€Z+`N"ѫ!2?Wd9bBg"\% Ii2. id e` f絎MӳbRoQe)Eh03 b`Ĺ Q".gJJiJiRd,9rD{G] z3v6?z#HWvԛz]ބ*&hphVRoqb[Ou4IiD1ٳu˥&NmQ,jȱ@XY\{:F&_XظZ4 LK1͠6G& In4FH,LI*5KcR( Bfؚ0y_fG#׎ d]Vkj)ǬHV`H (.`c9ʤBJR\ {ùٸ@5Jn\Ȯ\gTLL2ijjNu]\hh ݭF^ݦ:\G MO mM-sG'3 ZhI2W,R.2 ňr#* knČ6uEI4vIZ VЍr*ޥĎIAo_a*ȷPaVfu!=G- [+5ftcZx6 ʳG#0 :r@IJ![-永xqQ5Gʹ*&q+hz9k ʶ=!̋-R5εpU )/ 2$DyA!@”NvM:qSd,8̹/LZHUE(8:U)8#(rbʀ#)uJs|XR|tno2W*T%:!%jdO_׫/CIJ@)^Lѯ׷ J|Ur8bR\拊Aoon>nQV9"f52WËM:_TU3K1b0R}E_\mY%bbGПO"I s}Z<е7B]urn,Y1[W͠{Ar@5~ +keZ ˤ)Мq2iNJ 6ø,z2pRh{snT @ @ ҷc *jzיio(rF=ѝytA㧛3o C=q;NC0aB21WM~zfiWv4K鿯%^ֈuhL7x5{ݟJ>?8|V3}RʍȉPty@){Hi)PNuƤ 9hF&w+ps9Yv,IB8>v z\ _嚓'-ZvΣh<Ǭz"Dgf,jnE\PI+*r<,GW9aʹZTd F+EEqʨu4I}iD{ } 'NK#"rDcQr) DM)BzkQ\7Mˈ}L KnT:E8 矚z*ͤM-EO)'lA!p+p` n[V5hA@QNB"Rƥ K"ɰ` K-Goλ4L՟ѩRUp% XclLLS企'#aa+UXC^P$ooP&PM8&ȩTwn %ݑvi5L[4fc{<ٻVy B8ϝk.6.}^{&O*'ɓ&7{a\C KZlQ)|.b,Jw.ق\R{a~Q^_jkwdȮ(oBZ|K*/wS!qT#`7Z^`aW7mƈ/1e7D\X*EURr.*y]]fqb#SWɄt@Xk=-o]'W !hN8?aFcӪސ'"]T]jS!Uu3%O#/װ\~c<&%7GaLJtۇ'VH3j?i;{n-iua:ԃ7z$z?VJ\Z{<]^Ik6ZI5Y@tօudޑ&yq3%Nq)X hlcw(t1!d}.9P."2 m*[ 9MZBBZKBtI I'Dݗ#cmB$`=+L ~5#t5+mLIFeܨ$ą3SNDV;ud*YبPqepj5[NDĻ:2p‹\-vȍLK&5zm({mXMtqs#-ޓ8F?XDF7Z:U[6o3wM6_}s_j.Nhx97GT%]Ax)^/?͇S>?AQ+1n)N'm|Svsz'Cҏb vio=qLNTa;L?}uk[9ཕo toX_/Wߩz< y{M).#8i+m9+l-uҟMF\Ƥ/&~sbhdqv:LPzAe B,6A7>efؒKđK \́m? % %ˈm@Ȏ[R餔֎{%8W»ƙT'e/%n$\vIn/ٗE?ň3bP `墀8SH9KZJS긐OSO?W8W9R*KWGw@a8tɪ¯vw3߇e7pΑ bTosm*kM߉=SW;eЭUܞ*\^?}r?8j (WO;;4Fgdd칌l2;>A{lu7z1:ׅf3~]mo!~tzaz xV1IȒY4S>T ,b$42>y`o7;~N |׬7w RR(9Z7-rRpS޵ kձOy\@0 <ڝPs]zT|un?sg`?~M6 lN  oF^Po$>ۜAŵc(@ tT__#Qsu8Νumbk4;0* p;"1zIJHY*3:9'*gWv9n/mMh!V=`:8vj?vsƹHV^ގh%_d^]>%>]mfM6;{Wkz$n1tM⧏?qE8'N*XYzio_=}ruCmhՆ ;|)DGq׍#No\9) ;3k<9ӈޞeoh[mhV 7·v%0F%"+pv7$l NlN׼;Fp*J?6r]I6Ǎl 9q~Ů?mH`%BBJJ>;mP9%.{ct^:} 0;6YٯktKڮ i><{y0ݵ[^MQsҥkEKlY%AoX 'wRͲ38(a:Rm}h_:ʶw#rlkDo+DɗO^OQ$}ڠ5hPb*d)? r ڈU0QJZy~DE._&G$I")hE4M&ay&\,J;G>rZ8_kw|99E!iiގ /1J(*2 ~(Z6PGv1H5 S!)S 4aGA(R3BV/J8/bI/^-0$MTA? 6N+`hT z$e! RZ߯_*nE=i,}p8GI%YY9Q7P6:Wre9}R6Y[)3Ob_N~DBA2d}2XE2biྔXc^B1&^y"<,i!Ϻo1^^b0 f F5ӁБhx#t>j(uLc$L8cfTO!+s1I h]V7)aЎmQ0-juf;!-K:,xXWP9pv z`uX3eu:,/?p!(6Jrkf gܓOV\񫾠/S4ÏEE JM$ cgm93V$w~uz`Uf~'26ml&:e%,L0G<``QcNō雗}YahsG%^lzu/BT儧L^|"r"3筠qEw7֜7F5jk5荼 /~3q*ΪrJ"Ks*r [BP2 D>W}vQ\mjxK 2Z@W"ZdAà<8 q")XH/WQOS(Œ15ӋAK4Ybhy &ϵs֍ LכV0WbuɄJ!K!riVhYp΃P`8j,W'% 'U- a> 86ro8Y|P ^y8IJGAc%}x'BΙ3ecUց>xkz a Ţs-93 N84 XJ *W<`tZ+90vu kT4$DGy$"O27WɁ*^͗Mz2n)Վd2d"w/UYY֪M_'H "( 9#ސă!;>4{&c$#ϡRHE_W\\^L~ h5'I&W̪f!mD;~Ըupfh1wu,KQK@yY|IC<(uMzc92^jq6qGM<$Tt>2O}eWsRjc9:==QJ9sw!7fL?'d`u3XPJNEjZc@{E '+ he,H ;=^|;âx | h-LR@'H]AJ=xj2NGEOB6α0ZЩp'´%<s@ު96 JkҺ0S0甬L"wzh؎PSD7f!c3t }qYe>{@d2Ka"J9I7ƲƞD0#Ro 9fu@$4WdXi~+2ʓz3RL"T^V2|HyYCЛyɃnٺť6IǷkǫq}V8wN\h&)Ä,$\m܆ζ ys|oCrh<4,&k 2(4r|Ӑ}$3-(5gq޸2Cg7g$0Wj[@*2M78+m#Ibf6ɼ~h`vw?SEth^HJ"JmQU""̈w 0j}qǥb!AqNdak&B/@(Hd҃&MM}g=[ܺ(΍ʫOZpu r"Bm@,4</DXrFn_n *~9' w2j9,_DW bzĤmd3Ec6bN>88P{h$ăi<$.y(,iblUZ~EZf K&lT4S`|:S51!T *_&1[Fxu{lz0e0o{W8/Qs `c4B fa2'`럊d\.5!Qe|KRoDHeH\&1BA 25+pXKfݕλ|؍B!Y4y/ .J'W400]C(lg;ezPQOa6%KDĀvZ\!4_&ƂX5rp$j`|^eqH)! ,E#1d<{a3Ad%; 5zvo;M[\r~PqG]s}xV:j٢|WCi2m%_}q+Vqa 7˞G,է2Q3_`o=;O6XY6 [:[VoķhYǵKlQHO媁ޮrZ1.0*G&W(Fb1DdGR(*޶FJK:C/[j> S&-خZe<;`Ôc抋M(kiw|f|Kamp P̻KS rY9mB-ٲb5%67c+Z.Fu(/(:i?,t(*6N' MrM?Js:2i5 Gg8`T 6%eV FiEU %ʲE}@U >iE$ L0IU Dˍx,B mQX&h6;h#*rSDy[Oט)ïȷ~UHݫbo2oMFޯh5k< 7¿ fHAakHI>LqjA($C5  k6o/Ϭx߮x"+֥fI!x/Nrj&$n]0F$*XPdv)wqs(XNQE|HE ܯEpD.fñ8*teuK<,ӽXEG^Zÿ~+߿,Q|9ě".*Įf(a>?poo@r͐,̦wJQ*)w`)i_H!pvyy%c<ѧZSPx3LkZL` lxFJ`Hd<їTz׶V:V}2k*0VJ§T-mď&(s bX@2<!Dh5QBRQICqpIdjӚ-XL+g;bz["|O7R58njwJeBוSYÂ6/H7poӜ_oAb)cΐ!hJKe-C[6j_1mvz=v@D/20~Űl]S=Zѻ>Ƕi{Nj T9j?xOQRDpYq6LE*wB~:M1G _6d{ vϳ?pY*.V_3|kBrDg$3HZITgFOJҵ?V>'FW6fz!Cܪ!W?Pڵ]x~L\>H}m瞝' VѠB[Hչ#(YܠN "jDps/8~AD4O KIC&M8Z"h_}&s(]# \S;P(/{}H C>|$lѡ$V{2@]7e]Fffn"87mr[M<ߦ5F-{pw?Ƿ.!q2VV NU{2kird~XY<_og{BȒx}V|ϴ {=-cV8.am{4|nq'ʃi/?oh^qJݥ1CXXX. \ +F>f|yhg%J$Dz鬖$/.dJᓷӏTvkJ:;nJڮ[+5[F2{Oxj[h4#:M  ~6ּRNMֆEhRbbOnj6 |MW9aUsª9'I*{ڤ/- 8ڼ>޴],3s{yծ #7XB>k^0["6Ĺ-q4H3И|$8|bn&q0{oFb$ $G$)k8F}ԲHoTht( #E5dANoqİd0\b0[AsƵNɠkADtQ'!LW_U xM)! gQB$wĺ꺠uw$e7SHu |%"ydWJQ,M S3|}z$\R!$%S4UD.YTǔ;MMEbtg/-wbWI@cHyI|eʹvjr ΰOdp]8%(d B8eSH- V"Bw7O蚋@' ;bdȝ޷iqyNF~7s7hnȀ8,z ψB *jv- .TgsG82$w£+zjBd%9x|FahIZzEAA=kV"\igagY&P'Ӑ lb"!>029x?h%JH~.?&OBg@eRRn\2W9"?7Lr=DƬ,xO0BQFw?V ߠ(z(]-EּHHP)3V/6޾AqCk,;Xvu :h4K$@"oaW/e5ֻh0_H05HpE u(`rJy`ݱ~73wyJy 쇕~q~5è 7k ʴ=Gf/8[9Q; s(fCg 䫚Z Zn>P"e/U1Ofi(ť@qa E4E3R,lqF{ n XjzoZ',gKh?1x.`L49(c {.chMc]K1?r8?3m 98rMLĽWK^pD݂Sɏ"Rr3Z=nUӉY?}!JdbB4Wƣ2fB<2[ȗ Bªd̳`c]-Ίn9<[VXoIKKފ3W|{vOn å͚g*8Y~s0Wnu9{[8A|(ziޤ>1n:;'{$+X8K6װ3ݸ`EG.w ozܫy x=,zoXr)S7^@8nu$$IsO5ٯi(H4E(3Qh ѤH$$BVAcEZВ0 IJ\Ƨrq&y:OFE#|bgSy=Vu0"3= my%xB& Y&b4ŌLlvHH3xcGٟΎktOfs2"w/bZ5\qaaBdػwMŢ9˼_ȼ=$ I]IXs=ͽ.FddzY:76FS'C~z茗҃bQ{, Ɛ[_6`}F$`, llm4AVtd4$# *Y'9OWLW-qҪKp8\wNoJZOZ^PMاukXZv~v@!`)5~ @>7 yn`AM>ύ[F7;.NGzA)n\Khn|?UB]˃sM,AmC'FϔpSx"Bأmu^Itv~iJ'.f .ޟTt˟PH铎Xl4 <񾜊wB @kR  q>f'lY1 1ȖPXn8Zs)G-$Yv`A<r !D!y\qLRLyamj74cIJlmHWHViViGb(*|/%۷˩eËbIkMKa0>,V/!Os/$vK-njHޞ0jKzuUq79yeсny/TK< `7`=",odR.rB⠦ X"p@eJl6̞"~s#PeԌT0b'W|ݖi$C\SmvϿ0Ƙg JRKx"DX׺]. 㕱61δV"`1 K t!OBٚw\V5@ڼne=] J@u^y,)*:ي7`!57nlDwSmdMzu-裦^c)*;IQ\9KG@K+}dK:ʁf 7M4l}Z^`TX+{YNUI)gjܬ!c1E,1Q@P)nETEpcHWź6Jg?)s*FgrV9В0Jr&䲜+I-Ou TgT~kN a6~stԹn9ȷrkke룀](sPFt+Õ_GgO `inoOrG6V}/ ~~nhu"U~s\0l5[oinw_nc]7N/>YA;X_'ܞ( `si,ۮZe; o>R}$mdfص5{B{Hr-wtՙ_R6\9 u":&^.%qij#S\kޯXv>ҤyI~}f qfB3Aklqsqu71]9=ߙru;qĭ a?4-kh}߃xxȄ7bn+@o(ؒ 8h.xnj3(?W6"i0U^8"UI8g q88Շez=%;5=mxk({U{9A_$lN0vN\p:::A UV`,hd' Jt`/Av굵?Bd*Yɉ;<IIpS2U):(r qSPkyQں停q;!l:Ra۝Wq.a<4Jۃs<=/N1;׾?;Ӡ#oڸn_kMmYX=I .28v_m]FFz[WD=EyÛŷz0]e|VC-fSGAZ Y5f͟: .4.uOș^A=X2ʈGWu֠՗@ϑM5GӺ[ηFgAJ^i\UU-r6m6U %QΆ0'r׸cK ./ӍY>:Zbl$^bc!YoX*$|Y)\f)gr3f_}rbmjٓ|V/7yɻYy)y)*Alc ]GK,>$6rUc;`|+ ρ//A\%upo߆Xi'20$ࢃY| #Q8jw}⊪PycYX cT<y !)y S܇^KooaS}F6䰏aL+SI5F\]s,%bW51@>4piaELsFK?Y zV4#|RpΏqݏ_ywwuhCe9?7Z=O_ -zi0k; TҀ==RJj&8c#HkS3pfS=urw`;t5\z%v7 )+ĵeZmoذUt9_r&U$:\?\"f)j7ueY*]բ wԞպCJS kf5nN*S Ӫ䧺-Ok OSvӮ7IHsNtɃ"g^G%U$!Lkͽ:IdFҺ*1AhAdl*2,=\K47,8,skKPN)fBN@zwr;gi]ewY{펿4Gm~e6A!CJ6hY(gJʰ(@"t pQ9rvM.O_ |#bFs⾊X`* -IyI_JWdRJ-Xc\$?SvJ͋ujQ!oc5BI4Բ+TUƒ 7*-v`Ӕ(lAqS\$-|H#;l}6/#Y<=̢`orx|w05]O?=v?>vF5,􏋫e,`hFf;evwP̏5lyԛFYxwҰ)Cõ0&SMKԵ qGBJ|mz]'Pޗfὀl%Gn|iW̕\d&DBO'2 s|-Tm\ #^qdKC;esy ̔@{  龖}9)jw>}͐qKC03A 0]u̵6hoSApwJP.F)tb(Ӌyb=Q s2B6ecݩ׼[_NonN%{dwB&˦Fqy7䂍AFv,_ֿ%ӻ a!p]L8)#MO;z~i@L̂*b5`ܢ/Q}Jb+) |\eeasSܠL*բ9dh\Y)I-Fa/Q}JOVzV7"=y@pH+͛1c xj8~V*(J)= +g5J|[ʴR*F AJҖ$6{LVAoqV:AYiiJg4 b+DAglTŔ:TvI Rx/>Ƽ8̊+gx ^ZdI5lJoY&{jI2085YBjeX,>-gw:^5oa1sB @D6 %ôA-3YGWF$PA@k5e1Sth62)tu/tD/SP(GNȠVl:Iq&]>XFT+UPm+R$ %y+ğ,CsTz.6J<+=/"Gjf&Wj>B=~?ثx}Uw'/Ac/l=5 0i K{jb,lPbWI-M̹o  PJ&բm֝}P,P צV<+f{1D胚%Y8"Yڲam&epCA!/w pɺ}H戅?O nW{}|cЦXweop5W},7znyل3wʩqD8&Cv!v5-]HftR09@ W4kM` @Xac7(i%[|M f6ߵ BL+jze{Q7IU+^:(œnҍ&ݸS³0Ɛs q3 WӘIrYcML ƻFpǰY*MR&&¹|iJUFWJRr H|#r$ȥʃ)FA8{mV oTG &JvhcGNsJ4d˙eLhm94J:ҐHڼ`#1gWL9CVq5o;g0ĠpΪA ӵHGU&* 8Mvr()l L6B)2b2 [<8}X%7@Lr5&Rꄘf+颮?7ϩ )8n+TEEO0 +TTk'_zV g@Ga@yVZSďJ1sǺNb_S9M8n+)DRĘ:l=a]*"Qr?X|}˗]5ͩE,J%nm^vz}Ah<³)9#$9QlI֌+_%X3It?o|0Rd&STX Rz>b7^-[ZH\.4=tIG6亂NAtqHLGhcj ~=T1Ti'ɟbc~Q^>Upy}"7B&$OId7h\9K'N# (z#H*A* {dhIXB `̬Ki /S}F1c#?zɷبES}fed. rTQDsHpREqxU29>R\ [Nd QLMtP.%hBrowaqlwk`;.Ӌ<׸z8nN wY}2%Wuk%ɍĎ3x|NvZ}כcFR&+Zf_.sX}7{RhKJB 4ZTŸQiR\}wc:ri6yivMuONAY)#dP0TT+ը?YQZɴR3UqkDKNN{ &`=G-b2JU l 1R6>(G uePZZĿ߻ۉS`T⿗~RqF\.{wsp?_'aOEcgm#GW ,.Z᥊yXdŹ`g d8#,)v-YdՒ= Z캱XU^g$ئA-Ni-HԲ!oۿ8eoR,7wA``eL\y`򒎝E1N*n-KTQ*"Q]VoyT*n"~Ec*9x9k|E* ±>zˠΉ8Юeh ꃰw}Y3U̫JrɗIoC;mhg%ˍӚIrfN4^1WD=ǿNZej#) >ʙ~Ԋd Jw Xˑ]K[}':oB89ъ̮X&a*|gEdEdEdEɫh-Ap_Z2SA64mV j#FU @SDžvj7Wl=,w3 N4&ۂXaxY`r륋$| GyiVP=ncf daLz|p`4>x:J]/*pXjL3L'b9NcEHVUy^cʌL{d#E @ rxt qkGNkX1:Hyk瞚%)8P 82a ({J_ͯ,} sV9 fWX+í`S_}(&K-7.e͍%[ٙysp ̜')krE~]Z"d2oryw&IZu>h_<S~ ɴW|=NFF :'zB@gڥEd˴KYaK](|ʱg`qGԭ=ұBCg*kj8 4X-uQSAk&ԆQ)\46Hh*Ć1Lc}%oZp밫X)a1WBP'DUih`[-(ӮMX,{M90F@1";[2Ds:qc77ãUSc^= >"߳w|*;wp?uG$ÏgEdEdEdE1p:G7LPenb&/*= F*D+}yo غ7+I iuH=4G! D܀l6*KHA&V'!kRz$)oE4lը$>k6 bUUDhMž3ZՍkj"ԢAArPLXgD#j[,n 'p K&6t9j 5:" 7M Zv!&'Y-}IÄtAhj7hzj& yqRQJ"D5Nc2ZSfM N!6>YKi2\R [2瀒!.OT}ŤyR, ڲ{~jpP^DzHT$w4ˍ~=^lN7Uq^(AS^ ybOO Ɇiԁ_sl5;|=/"/"/"/LnɆp'ZBE  qA%ZkKr*WTҗw#q 7$ev,PaX#IYLt:=uAi`5Xh1BCM'd]$rA)[ X9IC,!!'u$LMzGWj b^rt4,!Ǜ<|$ {KӤ 47ee(bYA>vX*L@#bĹwӣf0K"rsT8DGAphId] JFӐKe)Fyi켔e%X⋞bo.&B`k񊰉ſW"noS, %4lA2{@ӄDHk2$~hXUt"(W i>K(jZywr< xwKʩͧ_E{Uu͛0x\@b; q}=7lZ%dnw@d7)rGZNIsN6ys%9/~_$OO` ?( @OFQayqL"(2xiԣso 4C"22I" {ECle{4SV& ybbAݪ{kC.jIhMq8(V#e]>z] Z>j@…ƘLo>g1̭f.ӺߝOo2;E4{w7F ̅v*tG1 q˷Kö`m;G?[W0bő?vF|좎VR,e^wqϽ) aadXKi*/.@;D%DS[v+A蔾#Gκ:}ڠ[c6KC?SIKaƑQxmو`G\~H% 1&OZي0%DrA)[tСͼ-` % LjUjpa;6%*H@jkV0gDe-ɏx1}| 虩5"d (<]/R7707-P3W`g*Z-%j?憷;39\}Mz^ũnK*ຮWMm wՖwܧaf+;u[q%S#u<b}<o@^i(=aC8%")|mY8뵕5 p)xcdmUܻyt yV&RA“~!*0[ S91`T(mj (P.uA# %e璀 x\LήA"a8#\O©f8^igi۟LWAzӫ4ոJNz8m hYF۽Æ*u\fGQ2Ý8[.yD ssή>ah̺SJ+C8DT0=\VxI#_s T`pTnngf:eܯ;)7TA|a^Z(Fe5ȺbEX++}S2eнސIxyo nnjaP\B)cecSWYOt=@7,F_v% }39 B֢_/4%" (;d?@x]UK@t;We"В 7[uQUUP0v卥-o,yFPC+l^[cU뛯1izxfխX]~QD94P!o0+p]FxZK&kRTj2M+r'Šˀ&8V/"oZpѩL!SٻrDf <^> * @a3ETaOO,?}93;8#uI&sUytƅ%^-ǂA;jUzh7U=ɢi Ś$ή FI8xw>yͤnAv9y8VV)PIgڊMd]cp<}ؘܱ; ˻uUkRDy|SRyRV: O)|.0QRAzǧfW0%1`gUZhe Z`wYq8tccD9ff+)mPBxW4R`z]OB?6-槱%/ٮv{N#)llpʆ,7Ϳb Rj\$ƨdreź9-:nxVt'ײ>w y}Sn&Z ֪B$ B HH3]RUH4jz1bW~z)j)(VD Ŗ2c1gq[^ǫv>r›тn:HK,L \4[pJL,6ʪ3oö/i rN[PM&  ^yQ9k-ï_Y䛏϶ȑb}@1X6NʛȾXwJT"[UR*U)TSZ?s9M7jgvޫ(jԽ@yķE75+In,Aޭ@XR>hт?ڗ\N7%,} *jVLK_ BM7MML8$ST9$Ց KpçcguC'C+_˱% Q %* 1)KYrՄ߿<0k@}&˘3O)8O}" WҶViNP/<&_1B[BkN G ;"t2) Yf% V}阦#s4Μ+.㲮;UDABMVIm~,[ȱARN7eb(MEXK pڟ"mıxH0wscSR5ER#-+v}Ѹ,>\'[>7?M}Ǹ]?lO)nZ@N lT$1p>c bmԱ8{e]}ةk((iU%DxI5,};ojЁf6uLj{uUz4sJd1#z&[t`6upzXO.A 'R<te]dP-U0^|5h`dsy'.IbpglHfTnrjƐcM'sEɓZʊ{C1a8hc%ӽܜ{]"N+ʷc1fhF O_ə & tj;LsLl:`Hc3i;bSH&UެB/[UN1:ou"va3+J;^?nc >WNA⭤9Є(GD2s1yceP 6Y=nI8KdgʼnbH.z1Lu,zaBSoRg4+|~zɶ59IDʬRFsH_/=a?> µ QˍZرe~Et|$22*}gIysirOtO#^nwxY;{Yv7|$l}?8S!ܛSP,U+GZxݱOg~i1`9 >O9Z!g阽&6ٞyTf<` wf6uǫ͢xp>xmɬ|`gaop y"1fr9U>Mp9\.z۹3m]^פmu<)5$9V!~ϱUpğKfb?.L0wζVmkynԚK fe~]RS#;&!}nr/~:`KzQvO6J~T&PZQQ1P|h92|27% ;փX^kTR;@@;M:@E̺O&Zkr1FTYƫhʘ<ւ6s4[`7Y>ԮoϏ5ߺ,Kj@ ƸN+\MbY!'ҹY5o&T\r8 cFuӺ0ƊȤLA*xYL\mhC@5V(Cj;@ O dp<^ NS0j3G1CĀ~^s?Zgaf Pj/S7z˻H$Ҩ5PeXW[ȷ^em^G+&63A= CP?ېG^8 ǍQRх'V Cf5nbP]VƐ5$#˒g|FP ;g,$jw.P)QL ꫤ7N5r!bVvxIꭞWUޏ>%I {m/&)`-NIcj6\0GMo5O[H*[ ]Y:rI+ s'M'yҏX2J V۰R7.uۻHr\mHc3YNog㙯|/jAn*sڎb&ͬjh.y؈I-WshW1Yգ) >)[b@b~&p?q7{jm?SFbvjJFBimQ{_ATXIsmYm|7h9aVzxеשּׁ}ފdb'q}<-zҝ~gU3Z@ΧyGͿ]~,%>>[E*9-PW! 0;3] K3Br I U<\>-85~Mw,|v2mdYm[kݜ/$+։ú^^mޞ3u|a ~w/|V}c=SHdU~ja:w 7t% E.n0?1rpAIdcQ1\34%0\Lf;fc Kbl*:Sd7>x9ēq⃽nkg?yKsQAQv1wB~:lMS//v]0;'}{ *E!L藃m/?m#+(fwPR6$vqs5,Ƽ _HEʠJb$0_wOn& LC"&fYt](f.}ϻV 9^.|e+*|\礓{sa5CsH]t$G=zҐ!mBY=nqx?A7o.%Se)5@(hSk0Q _7Z׃Oy.ՌRLtSR.-TǁƁ#M{;H c"jL65L Ass+R?~`,&P]6eށaE·ÏHr*>se0, yGhPkQ3]5mn>X:x z;@ZB' $u?;.jp`HtlK؁C$hXݚy $#a^GZ+a1nĂGsܺ+JŹ2{F(jխE@{:RW1(3܅gZm k-jJq]dW`ŻLʉӱႏ;GCedɐKTLyŏVYԄlژ&{جP"Tt2H>t `\h@M[#qdG+,jZ'jR9x5Jd]YٽB]9DF`Yb8@=<{l`q<>W.XyԽ8g|{:>m7-b6NFI8޷Rj !wkPMG&[Y^6iWtaȨZ`9rpy $-Kp:pgv9'Z'o6֔cKG±"Q Juw|`r`IkpbΙ} }k6QVLs'sJo/p}z 0_Rp|>'Kt2D~\@Hy Y|աGٞthڙ;jiƜKL1 I/w*j.09=)RY}qE+>E%)SPדͅWp[J|*bIdj,Go$Z?t'!H`nUv&DmN~qC r=ӂEAua}(Z!z7?-@r+g.5 8~q@D$E?cz)^S5]ُylF;{S[^*e~mo Wޝ2J߈*LҮ,;Œz;9@8!)X=H+-)TQ CM$ خ89!ɺtKjįp&8l{UW~zΡE'7B-&LU&|ʯg?}_V5'X0;3%:.ʫ^((f[}LV!:%cX4JZ;?zGg ޴5jU&R_'ߩ3UquJ crQߩ;KOC1C eQTrd DkWP"[x{! q.8;R.lm#@Rp4NJl\ c+̗ʄ)ϣA' Q Q\s81 EGM b#<ߜB %42`әj[L;]H(|$ξAEY HWj>f*Jq1ԉD1c,'m,58Z fHr5@eR{(3m;5JOmF/~E ZqbbƉ A+cdԂZ8{p47M&?D)!4B<^ @m2D7uފP`!Ć[B[$=b0/Pb#'QפncAV A#OVMlA-1#x!H0稜 \r _HÔB)p)ˑͩE&b%S.M-ij~.h)Z#Fq,_G]VKպ1KOvaZS@*2SK"vVE,NӜp%7!hxqALEw5%b*$tН\V48h&`2 H"@rw{-8.>x)xR\ޖ߾ ZJEegV(VGU7^+MH1J>|>v,:NWB"MMS=2֣͑Y6afN}̮%:\#\KMoV̞s%6:[`Sx`kVha 3y ,|.)^~xVrxY ZV8mhA P}2yې K^)>AVa*5ok&72n$l6$Ĝ6R=*('1,g!< >ANpz$m<#5?ZdMYe;K^k6ۖFx8xhZ~Kq(iHv+f47{w)+j2ʊFt0 b*nyz><{[&Hʹ_th47%zgkjw@ "t{,Uev 7>PC$3?@$R wU {/gk]ѷBpru?޹~ u߆*8b16r{Z5jc跪#;L!*iJ5䆡tD =]p!bFĹsDlg"ܻ70Ftr=V=T[i;TOZ̹JG[$[R? y cΜ&%Z[[LWjVD޹lPZ#_~~Cc 5W7f^Z&΋a?ogiW<±(NZR'T?7.ռfҫy3 Rߕ?<`Z`g?{F#_\+|A&7p>YGoQՒlcgEm*X iAҏpT䩔t=e6`Lu[/[fFz}r;IB^VTg쩞;ANAb#:(n^ns3EJ:N@nSnN;J|NO⩜NO:ncH ўLe6fA6ܪ߬$qzl/gW뿧;Rب0)wѾ}&/"ݽ7F( !ʞCTQ!{Q!/Ԡ|{nv3bIQ؛ڝ:Gz"?ZƫqmOORi)1jY\JmWX}+# aq>{rol=/п)vykInʫxu4{s{c+H/*3JIZRF 6Rá`sRDo+ޘSzT(( ӛ_Piȕ`)$R#UV ,]D\ꈮEۦVlcZ!ynA6V9$%yGrzX[PA١ l+R?<jͲ.qˆc 95Q['ЧxwkEJFɸL |O-YohO闟f77?g[.[xrneS0S5'Sʗ%*ѹX/b %c.:]p)p)EjUGڞyT,B1x暖z7S&+XW_ҀMNwI_)mۻم[$spKOPmU(y.UE8tzujU$HaJ7/?ybF=yB->_!:]^]W_mRK?b~ʿC4c#R)ihZV;ўBSu%*}nI~Qm?V h~`> / ^y3ot>O;Ĥ6?5B\%$Իh_kkcstȤdԳ3qݾ:Yw[@lTs^;R`tL]e4,!(Э2Pc>,G@o {!MN0:cR4Fj xKh H(G993}<邁6ZjMAupEQ.e< LKt&5sPH0YWrx}j䴚WT;^)%NzdrC2\?G;vU`Q݄x#ԍ44, %]2GF[5hޢZF!0B B:|<_^{F@cp~!j6_nV͹]} o>R9= QbXb5=3\ YBNqA5G4n;ppF1GQE9gEOI$Apøj5TRhrX\X[uE? Rv_$5QpOij(P3%g!aRǬ<,eAf8dU!0ń`j!db D0D@Bh8h&F(,EPm0 ^v+S*=.^a]V9T|t23 PvpVijY4C%F--e#IGDNG.(hp؄ %Jc砺/_-@ɑk'ʋ5Dy&:_3q-LQ MG)~Zh2DFB%jTK^pP(΋oec T2gV9q`?Ü6D mpBJkA@W0fLW f|ڏdQ͑l ȴ]9ڮ=Dv3ş~`duhO~wfc}/7[D7qMks0VL9[;5B/Aq|9^ QeWqC灖{{/WB*TgU* 1hnh~Zq7#A{Kxxv0u 72bOi)%nQ*Z N3i.S!}_BAD;/0nQ;Fe+yW86c~3 }/8]fj]͵y+븻y8kCH\A\jjpF^Te}ϣ Ƙ]HNӍg: &Zn۳+@wϕ4QSk@`gk\u;Y1jMH(J<T+E)۪ gQ;JmͷO;ml=ozƐ.52;yRSnN;Jۜ笖FMSO3E'S<<5@w׌4"yiDӷi .LհNpnVuhjڳ|M5:b8S[d[\c@:7ZXUdwż` 5Ŝf*n̫*/.L3P% Z({UFtN":9p(SxJi(_Sބ8`}-9w'3z/3X?p)sʮvӝI3jZ rD%mFyKEגƐ.52՝!uvS N"ѩ}GvхBN;ej1$䅋hW(z92["y5R՟|qb?)aFՠ|Gj$TvmȪ-1F nRʼn3lxt߽Ӟ{OENk:fKÒΈæḇ#)fNtHN'F\r8ώE^,vN%!L8U;| >ՓF1!9 (n'#%64)zÕ$\)1)0LIbΞ*g.|[4,xKeK-ʰl"20Q̛؉%`G$L=.S }{m,<( zNz41tI^pBٶ5U39boR:ƨhKp:(% J:7iI}kU;nM=m)@iamLdaTD5Uaسpg0#Q YpĬ%-qx2 (F%[߇<_pEG{Le"pS:v̦K q:p Niwx7B7싑P,8P؍ѷ'|D8ȝҥXWmW XK-ܛh*`}jt1mRH?M*;m}WY8J4eQ Q4ضf tg^|`y6¤"Tx#!k5~Tq `Yguׇ!\ɤ@lL6O}ĭxqx>]}n]}\'q ֓w/pãv׉v݀-GB`>ܰu;j~O̼Z,eH6]O"s\O(=18:~Ӻøt\= }\m9vY K-e |N Y.[x\lya)S8s0e%>^ƹW$Gy _HFl0' TxO1h"^]I~Co_)-olRmMԛ9C6~OE-WuDsuT E˫U5LI`>nE5U,kR1m4+i&Vburs^aαUWiD]^y4G+oF2zvQ3B3 MoK&Yx92ӟ:S?skDgsW{׶ C%ؗڐD K<($h$.[Kd%c[m68>87_jPxgg%^6;#H)-nwmḞ qϏЮ Vԛ;m|_n} e\ekkWTKR+r)7O*Ǟ4e WD1﯂G ~۳0C5v"N3.C7m|?;HnSR-c3)w %CZjxͦ/`e([QJP)LW^ JЗڵ/uڞJЗ& [ɗR6~EWO~~ҍ^ H=7O> Z䑞BjyM=49<-KG,'xwzeO3 l܏٢Gy'q7VEPpIjEgMS U8ݶK~nm s.E6b#T<5[j!ʻpyHp>yBu}•ǥݗӽ} (` тm6&pm-t%iZպ` [1$_UM#m16EdbPԖL'^{hv d$3ANSK"oEq@Cu!CqUeBi#n:qB%j8ց-_wuR|CY?ArH4y[N5n9͢Ŝp{Y4POis45Yw݌78g.}7oi18ʗ Jo+{Wq ;}3zbN.Âq,bH%Wo`8Y Ϝկp.+var/home/core/zuul-output/logs/kubelet.log0000644000000000000000004623172615154750330017713 0ustar rootrootMar 13 07:38:59 crc systemd[1]: Starting Kubernetes Kubelet... Mar 13 07:38:59 crc restorecon[4696]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:38:59 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 07:39:00 crc restorecon[4696]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 13 07:39:00 crc kubenswrapper[4876]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 07:39:00 crc kubenswrapper[4876]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 13 07:39:00 crc kubenswrapper[4876]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 07:39:00 crc kubenswrapper[4876]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 07:39:00 crc kubenswrapper[4876]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 13 07:39:00 crc kubenswrapper[4876]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.776506 4876 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.783881 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.784334 4876 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.784447 4876 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.784578 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.784679 4876 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.784775 4876 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.784881 4876 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.784993 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785091 4876 feature_gate.go:330] unrecognized feature gate: Example Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785189 4876 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785338 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785442 4876 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785564 4876 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785674 4876 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785773 4876 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785869 4876 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.785975 4876 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786091 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786190 4876 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786317 4876 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786420 4876 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786534 4876 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786642 4876 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786738 4876 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786845 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.786947 4876 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787047 4876 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787142 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787278 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787383 4876 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787480 4876 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787613 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787718 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787827 4876 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.787928 4876 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788023 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788118 4876 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788223 4876 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788375 4876 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788485 4876 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788582 4876 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788688 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788787 4876 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788883 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.788989 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.789095 4876 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.789193 4876 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.789348 4876 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.789487 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.789606 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.789704 4876 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.789800 4876 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.789904 4876 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790011 4876 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790114 4876 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790215 4876 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790362 4876 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790464 4876 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790560 4876 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790679 4876 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790780 4876 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790877 4876 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.790974 4876 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.791074 4876 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.791184 4876 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.791318 4876 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.791438 4876 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.791541 4876 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.791639 4876 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.791739 4876 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.791838 4876 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.792113 4876 flags.go:64] FLAG: --address="0.0.0.0" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.792272 4876 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.792411 4876 flags.go:64] FLAG: --anonymous-auth="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.792518 4876 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.792622 4876 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.792721 4876 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.792825 4876 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.792926 4876 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.793025 4876 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.793311 4876 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.793423 4876 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.793549 4876 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.793654 4876 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.793752 4876 flags.go:64] FLAG: --cgroup-root="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.793851 4876 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.793951 4876 flags.go:64] FLAG: --client-ca-file="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794067 4876 flags.go:64] FLAG: --cloud-config="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794171 4876 flags.go:64] FLAG: --cloud-provider="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794310 4876 flags.go:64] FLAG: --cluster-dns="[]" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794437 4876 flags.go:64] FLAG: --cluster-domain="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794541 4876 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794645 4876 flags.go:64] FLAG: --config-dir="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794743 4876 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794854 4876 flags.go:64] FLAG: --container-log-max-files="5" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.794958 4876 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795058 4876 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795158 4876 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795289 4876 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795395 4876 flags.go:64] FLAG: --contention-profiling="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795551 4876 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795676 4876 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795781 4876 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795880 4876 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.795984 4876 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796084 4876 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796183 4876 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796325 4876 flags.go:64] FLAG: --enable-load-reader="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796448 4876 flags.go:64] FLAG: --enable-server="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796552 4876 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796663 4876 flags.go:64] FLAG: --event-burst="100" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796783 4876 flags.go:64] FLAG: --event-qps="50" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796893 4876 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.796996 4876 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.797095 4876 flags.go:64] FLAG: --eviction-hard="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.797199 4876 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.797332 4876 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.797493 4876 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.797624 4876 flags.go:64] FLAG: --eviction-soft="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.797737 4876 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.797845 4876 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.797946 4876 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798046 4876 flags.go:64] FLAG: --experimental-mounter-path="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798145 4876 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798282 4876 flags.go:64] FLAG: --fail-swap-on="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798391 4876 flags.go:64] FLAG: --feature-gates="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798496 4876 flags.go:64] FLAG: --file-check-frequency="20s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798596 4876 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798695 4876 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798794 4876 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798893 4876 flags.go:64] FLAG: --healthz-port="10248" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.798994 4876 flags.go:64] FLAG: --help="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.799109 4876 flags.go:64] FLAG: --hostname-override="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.799211 4876 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.799346 4876 flags.go:64] FLAG: --http-check-frequency="20s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.799449 4876 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.799549 4876 flags.go:64] FLAG: --image-credential-provider-config="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.799651 4876 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800105 4876 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800338 4876 flags.go:64] FLAG: --image-service-endpoint="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800395 4876 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800443 4876 flags.go:64] FLAG: --kube-api-burst="100" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800490 4876 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800537 4876 flags.go:64] FLAG: --kube-api-qps="50" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800584 4876 flags.go:64] FLAG: --kube-reserved="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800632 4876 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800679 4876 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800729 4876 flags.go:64] FLAG: --kubelet-cgroups="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800772 4876 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800822 4876 flags.go:64] FLAG: --lock-file="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800870 4876 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800915 4876 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.800960 4876 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801037 4876 flags.go:64] FLAG: --log-json-split-stream="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801087 4876 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801131 4876 flags.go:64] FLAG: --log-text-split-stream="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801172 4876 flags.go:64] FLAG: --logging-format="text" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801219 4876 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801388 4876 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801433 4876 flags.go:64] FLAG: --manifest-url="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801513 4876 flags.go:64] FLAG: --manifest-url-header="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801567 4876 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801610 4876 flags.go:64] FLAG: --max-open-files="1000000" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801655 4876 flags.go:64] FLAG: --max-pods="110" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801698 4876 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801741 4876 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801783 4876 flags.go:64] FLAG: --memory-manager-policy="None" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801826 4876 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801878 4876 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801923 4876 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.801970 4876 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802032 4876 flags.go:64] FLAG: --node-status-max-images="50" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802074 4876 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802117 4876 flags.go:64] FLAG: --oom-score-adj="-999" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802160 4876 flags.go:64] FLAG: --pod-cidr="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802210 4876 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802287 4876 flags.go:64] FLAG: --pod-manifest-path="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802346 4876 flags.go:64] FLAG: --pod-max-pids="-1" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802391 4876 flags.go:64] FLAG: --pods-per-core="0" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802439 4876 flags.go:64] FLAG: --port="10250" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802486 4876 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802530 4876 flags.go:64] FLAG: --provider-id="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802577 4876 flags.go:64] FLAG: --qos-reserved="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802621 4876 flags.go:64] FLAG: --read-only-port="10255" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802668 4876 flags.go:64] FLAG: --register-node="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802710 4876 flags.go:64] FLAG: --register-schedulable="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802753 4876 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802804 4876 flags.go:64] FLAG: --registry-burst="10" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.802851 4876 flags.go:64] FLAG: --registry-qps="5" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803077 4876 flags.go:64] FLAG: --reserved-cpus="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803126 4876 flags.go:64] FLAG: --reserved-memory="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803172 4876 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803215 4876 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803272 4876 flags.go:64] FLAG: --rotate-certificates="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803320 4876 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803365 4876 flags.go:64] FLAG: --runonce="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803415 4876 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803460 4876 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803507 4876 flags.go:64] FLAG: --seccomp-default="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803551 4876 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803593 4876 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803636 4876 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803679 4876 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803726 4876 flags.go:64] FLAG: --storage-driver-password="root" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803769 4876 flags.go:64] FLAG: --storage-driver-secure="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803813 4876 flags.go:64] FLAG: --storage-driver-table="stats" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803855 4876 flags.go:64] FLAG: --storage-driver-user="root" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803903 4876 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803947 4876 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.803991 4876 flags.go:64] FLAG: --system-cgroups="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804040 4876 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804088 4876 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804130 4876 flags.go:64] FLAG: --tls-cert-file="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804173 4876 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804225 4876 flags.go:64] FLAG: --tls-min-version="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804292 4876 flags.go:64] FLAG: --tls-private-key-file="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804335 4876 flags.go:64] FLAG: --topology-manager-policy="none" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804386 4876 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804434 4876 flags.go:64] FLAG: --topology-manager-scope="container" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804477 4876 flags.go:64] FLAG: --v="2" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804522 4876 flags.go:64] FLAG: --version="false" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804567 4876 flags.go:64] FLAG: --vmodule="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804610 4876 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.804652 4876 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.804887 4876 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.804939 4876 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.804985 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805028 4876 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805071 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805112 4876 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805153 4876 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805195 4876 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805262 4876 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805307 4876 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805350 4876 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805393 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805434 4876 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805528 4876 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805615 4876 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805661 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805704 4876 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805745 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805792 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805843 4876 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805890 4876 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805936 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.805984 4876 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806027 4876 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806070 4876 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806111 4876 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806154 4876 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806197 4876 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806267 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806315 4876 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806358 4876 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806401 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806447 4876 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806489 4876 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806531 4876 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806581 4876 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806624 4876 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806666 4876 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806709 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806752 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806794 4876 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806835 4876 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806882 4876 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806925 4876 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.806966 4876 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807008 4876 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807049 4876 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807090 4876 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807131 4876 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807177 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807220 4876 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807279 4876 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807331 4876 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807376 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807418 4876 feature_gate.go:330] unrecognized feature gate: Example Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807460 4876 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807507 4876 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807549 4876 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807591 4876 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807640 4876 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807686 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807730 4876 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807774 4876 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807895 4876 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807940 4876 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.807982 4876 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.808029 4876 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.808072 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.808119 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.808161 4876 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.808209 4876 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.809090 4876 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.819961 4876 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.820011 4876 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820145 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820167 4876 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820178 4876 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820188 4876 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820196 4876 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820204 4876 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820212 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820220 4876 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820228 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820258 4876 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820269 4876 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820280 4876 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820290 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820299 4876 feature_gate.go:330] unrecognized feature gate: Example Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820307 4876 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820315 4876 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820324 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820334 4876 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820343 4876 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820352 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820361 4876 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820369 4876 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820379 4876 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820387 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820394 4876 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820403 4876 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820411 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820419 4876 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820427 4876 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820436 4876 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820444 4876 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820452 4876 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820459 4876 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820468 4876 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820478 4876 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820485 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820493 4876 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820501 4876 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820509 4876 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820516 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820524 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820532 4876 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820540 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820548 4876 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820556 4876 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820563 4876 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820571 4876 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820579 4876 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820587 4876 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820594 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820605 4876 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820616 4876 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820625 4876 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820633 4876 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820642 4876 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820650 4876 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820658 4876 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820665 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820673 4876 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820681 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820688 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820697 4876 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820705 4876 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820739 4876 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820750 4876 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820759 4876 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820769 4876 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820778 4876 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820786 4876 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820795 4876 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.820805 4876 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.820819 4876 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821074 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821088 4876 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821097 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821106 4876 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821115 4876 feature_gate.go:330] unrecognized feature gate: Example Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821125 4876 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821136 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821146 4876 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821154 4876 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821162 4876 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821169 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821177 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821185 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821193 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821200 4876 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821208 4876 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821216 4876 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821223 4876 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821231 4876 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821262 4876 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821269 4876 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821277 4876 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821287 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821295 4876 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821303 4876 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821312 4876 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821319 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821327 4876 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821336 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821346 4876 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821356 4876 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821365 4876 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821374 4876 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821383 4876 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821395 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821403 4876 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821411 4876 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821419 4876 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821427 4876 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821435 4876 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821443 4876 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821450 4876 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821458 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821466 4876 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821473 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821503 4876 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821513 4876 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821522 4876 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821531 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821543 4876 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821552 4876 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821560 4876 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821567 4876 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821575 4876 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821586 4876 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821596 4876 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821605 4876 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821613 4876 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821620 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821628 4876 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821638 4876 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821647 4876 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821655 4876 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821662 4876 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821670 4876 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821678 4876 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821686 4876 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821694 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821702 4876 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821709 4876 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.821718 4876 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.821730 4876 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.822037 4876 server.go:940] "Client rotation is on, will bootstrap in background" Mar 13 07:39:00 crc kubenswrapper[4876]: E0313 07:39:00.826702 4876 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.830024 4876 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.830123 4876 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.831536 4876 server.go:997] "Starting client certificate rotation" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.831564 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.832720 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.857842 4876 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.860394 4876 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 07:39:00 crc kubenswrapper[4876]: E0313 07:39:00.861982 4876 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.883951 4876 log.go:25] "Validated CRI v1 runtime API" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.928305 4876 log.go:25] "Validated CRI v1 image API" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.930647 4876 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.937620 4876 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-13-07-34-42-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.937651 4876 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.953267 4876 manager.go:217] Machine: {Timestamp:2026-03-13 07:39:00.950410916 +0000 UTC m=+0.621189918 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:b1858cc2-71ca-4a24-ba44-14334b24f2dd BootID:b24a1e5a-8f3e-43aa-b425-2a9854971c62 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:64:47:d3 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:64:47:d3 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e3:cd:fa Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9b:02:87 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:39:33:68 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:38:6b:2e Speed:-1 Mtu:1496} {Name:eth10 MacAddress:56:d4:37:e1:c5:5b Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3e:ce:64:ab:dd:1b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.953470 4876 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.953623 4876 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.954616 4876 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.954794 4876 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.954830 4876 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.955042 4876 topology_manager.go:138] "Creating topology manager with none policy" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.955052 4876 container_manager_linux.go:303] "Creating device plugin manager" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.955598 4876 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.955631 4876 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.955853 4876 state_mem.go:36] "Initialized new in-memory state store" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.956169 4876 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.959436 4876 kubelet.go:418] "Attempting to sync node with API server" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.959460 4876 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.959485 4876 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.959502 4876 kubelet.go:324] "Adding apiserver pod source" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.959516 4876 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.963613 4876 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.964592 4876 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.966536 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:00 crc kubenswrapper[4876]: E0313 07:39:00.966688 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.966674 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:00 crc kubenswrapper[4876]: E0313 07:39:00.966808 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.967220 4876 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969698 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969732 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969742 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969753 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969768 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969781 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969790 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969825 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969903 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969920 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969955 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.969965 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.972093 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.972872 4876 server.go:1280] "Started kubelet" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.975113 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.975146 4876 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 13 07:39:00 crc systemd[1]: Started Kubernetes Kubelet. Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.975616 4876 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.981658 4876 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.981718 4876 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 13 07:39:00 crc kubenswrapper[4876]: E0313 07:39:00.978500 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.983114 4876 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.974776 4876 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.983278 4876 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.988320 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:00 crc kubenswrapper[4876]: E0313 07:39:00.988553 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="200ms" Mar 13 07:39:00 crc kubenswrapper[4876]: W0313 07:39:00.988915 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:00 crc kubenswrapper[4876]: E0313 07:39:00.989024 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.989215 4876 factory.go:55] Registering systemd factory Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.989754 4876 factory.go:221] Registration of the systemd container factory successfully Mar 13 07:39:00 crc kubenswrapper[4876]: E0313 07:39:00.988959 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.39:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189c56929cc9f0be default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,LastTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.989768 4876 server.go:460] "Adding debug handlers to kubelet server" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.990778 4876 factory.go:153] Registering CRI-O factory Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.990818 4876 factory.go:221] Registration of the crio container factory successfully Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.990893 4876 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.990920 4876 factory.go:103] Registering Raw factory Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.990944 4876 manager.go:1196] Started watching for new ooms in manager Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.991647 4876 manager.go:319] Starting recovery of all containers Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996533 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996590 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996610 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996624 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996637 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996652 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996665 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996678 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996692 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996705 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996719 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996731 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996745 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996760 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996773 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996787 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996799 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996812 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996825 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996838 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996853 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996866 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996879 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996892 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996907 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996920 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996936 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996950 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996964 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.996978 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997002 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997018 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997031 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997044 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997058 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997071 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997098 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997111 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997123 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997136 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997151 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997188 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997201 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997213 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997225 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997267 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997282 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997296 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997310 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997323 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997336 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997349 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997368 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997392 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997422 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997440 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997480 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997496 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997510 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997523 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997535 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997547 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997560 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997601 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997615 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997629 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997642 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997654 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997668 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997682 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997694 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997710 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997729 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997755 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997786 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997802 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997833 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997845 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997860 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997873 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997885 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997904 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997917 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997931 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997945 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997960 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997975 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.997989 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998003 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998016 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998031 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998043 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998057 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998071 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998086 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998121 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998135 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998150 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998169 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998185 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998201 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998215 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998230 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998269 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 13 07:39:00 crc kubenswrapper[4876]: I0313 07:39:00.998295 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998317 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998332 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998345 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998361 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998377 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998398 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998418 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998433 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998448 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998488 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998602 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998624 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998644 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998682 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998723 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998742 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998799 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998814 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:00.998827 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.002910 4876 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.002950 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.002969 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003002 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003014 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003028 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003040 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003052 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003098 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003113 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003124 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003153 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003165 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003177 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003191 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003257 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003288 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003300 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003328 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003360 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003374 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003387 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003401 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003431 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003444 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003472 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003516 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003544 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003611 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003625 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003638 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003650 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003682 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003702 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003714 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003753 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003765 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003777 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003790 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003802 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003814 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003826 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003839 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003935 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003949 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.003962 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004022 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004037 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004069 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004083 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004095 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004165 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004184 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004198 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004210 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004257 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004270 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004301 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004313 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004344 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004356 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004368 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004416 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004428 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004456 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004470 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004482 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004525 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004538 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004549 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004578 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004624 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004636 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004647 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004659 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004726 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004739 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004751 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004763 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004817 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004850 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004863 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004876 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004920 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004932 4876 reconstruct.go:97] "Volume reconstruction finished" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.004941 4876 reconciler.go:26] "Reconciler: start to sync state" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.022725 4876 manager.go:324] Recovery completed Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.031715 4876 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.034035 4876 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.034120 4876 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.034209 4876 kubelet.go:2335] "Starting kubelet main sync loop" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.034380 4876 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 13 07:39:01 crc kubenswrapper[4876]: W0313 07:39:01.038612 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.038724 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.041207 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.043008 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.043042 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.043052 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.044042 4876 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.044061 4876 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.044080 4876 state_mem.go:36] "Initialized new in-memory state store" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.065923 4876 policy_none.go:49] "None policy: Start" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.066812 4876 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.066839 4876 state_mem.go:35] "Initializing new in-memory state store" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.083355 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.134053 4876 manager.go:334] "Starting Device Plugin manager" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.134142 4876 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.134162 4876 server.go:79] "Starting device plugin registration server" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.134480 4876 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.134942 4876 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.134969 4876 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.135727 4876 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.135994 4876 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.136027 4876 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.145380 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.189915 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="400ms" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.235563 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.236886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.236925 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.236937 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.236967 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.237462 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.335361 4876 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.335552 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.337189 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.337255 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.337285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.337451 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.337977 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.338022 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.338619 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.338660 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.338671 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.338807 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.338944 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.338986 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.339401 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.339455 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.339470 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.339844 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.339879 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.339894 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.340140 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.340170 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.340184 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.340362 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.340542 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.340632 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341334 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341422 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341459 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341508 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341525 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341481 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341739 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341891 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.341942 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.342694 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.342733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.342749 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.342757 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.342851 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.342928 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.343203 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.343282 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.344206 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.344300 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.344324 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.411972 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.412224 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.412441 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.412591 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.412731 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.412916 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.413060 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.413215 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.413422 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.413564 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.413719 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.413894 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.414166 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.414333 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.421260 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.438430 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.439897 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.439961 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.439981 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.440023 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.440699 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524466 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524568 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524600 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524660 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524717 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524796 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524826 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524882 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524910 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524932 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524976 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.524999 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525046 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525074 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525294 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525403 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525424 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525462 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525475 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525469 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525498 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525517 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525507 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525489 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525528 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525546 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525572 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525575 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.525876 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.591226 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="800ms" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.685760 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.695098 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.727565 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: W0313 07:39:01.731594 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-38d54abca19f44c257d06d5c5342291cdcc943bba87bf9d9102f25d3dceec27f WatchSource:0}: Error finding container 38d54abca19f44c257d06d5c5342291cdcc943bba87bf9d9102f25d3dceec27f: Status 404 returned error can't find the container with id 38d54abca19f44c257d06d5c5342291cdcc943bba87bf9d9102f25d3dceec27f Mar 13 07:39:01 crc kubenswrapper[4876]: W0313 07:39:01.739710 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-4bc924cf9085ea457041e0abd276d1f717896aa399ef81ea0fe7157ac060384a WatchSource:0}: Error finding container 4bc924cf9085ea457041e0abd276d1f717896aa399ef81ea0fe7157ac060384a: Status 404 returned error can't find the container with id 4bc924cf9085ea457041e0abd276d1f717896aa399ef81ea0fe7157ac060384a Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.746857 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: W0313 07:39:01.749691 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-3e52dadb5073606e9f91b536068c1df7d835b486e6214976cc1b4c6e55e2ab57 WatchSource:0}: Error finding container 3e52dadb5073606e9f91b536068c1df7d835b486e6214976cc1b4c6e55e2ab57: Status 404 returned error can't find the container with id 3e52dadb5073606e9f91b536068c1df7d835b486e6214976cc1b4c6e55e2ab57 Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.755401 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:01 crc kubenswrapper[4876]: W0313 07:39:01.764599 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f0a97431f2b6502df0ea47e3bab228e0d3cd88121764b3639ff501ddbef06d01 WatchSource:0}: Error finding container f0a97431f2b6502df0ea47e3bab228e0d3cd88121764b3639ff501ddbef06d01: Status 404 returned error can't find the container with id f0a97431f2b6502df0ea47e3bab228e0d3cd88121764b3639ff501ddbef06d01 Mar 13 07:39:01 crc kubenswrapper[4876]: W0313 07:39:01.775725 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-289a791fbaa0769311fea18aefd8122832682d25a56a30b24984b69202ced2db WatchSource:0}: Error finding container 289a791fbaa0769311fea18aefd8122832682d25a56a30b24984b69202ced2db: Status 404 returned error can't find the container with id 289a791fbaa0769311fea18aefd8122832682d25a56a30b24984b69202ced2db Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.841766 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.842899 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.842945 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.842960 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.842994 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:01 crc kubenswrapper[4876]: E0313 07:39:01.843532 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Mar 13 07:39:01 crc kubenswrapper[4876]: I0313 07:39:01.989383 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.042859 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f0a97431f2b6502df0ea47e3bab228e0d3cd88121764b3639ff501ddbef06d01"} Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.044381 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3e52dadb5073606e9f91b536068c1df7d835b486e6214976cc1b4c6e55e2ab57"} Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.045590 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"38d54abca19f44c257d06d5c5342291cdcc943bba87bf9d9102f25d3dceec27f"} Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.047706 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4bc924cf9085ea457041e0abd276d1f717896aa399ef81ea0fe7157ac060384a"} Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.051349 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"289a791fbaa0769311fea18aefd8122832682d25a56a30b24984b69202ced2db"} Mar 13 07:39:02 crc kubenswrapper[4876]: W0313 07:39:02.088887 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:02 crc kubenswrapper[4876]: E0313 07:39:02.088993 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:02 crc kubenswrapper[4876]: W0313 07:39:02.150093 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:02 crc kubenswrapper[4876]: E0313 07:39:02.150264 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:02 crc kubenswrapper[4876]: W0313 07:39:02.185765 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:02 crc kubenswrapper[4876]: E0313 07:39:02.185878 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:02 crc kubenswrapper[4876]: W0313 07:39:02.194613 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:02 crc kubenswrapper[4876]: E0313 07:39:02.194697 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:02 crc kubenswrapper[4876]: E0313 07:39:02.392564 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="1.6s" Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.644373 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.647297 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.647364 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.647383 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.647417 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:02 crc kubenswrapper[4876]: E0313 07:39:02.647955 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.913667 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 07:39:02 crc kubenswrapper[4876]: E0313 07:39:02.914587 4876 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:02 crc kubenswrapper[4876]: I0313 07:39:02.989384 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.054989 4876 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f" exitCode=0 Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.055054 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f"} Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.055125 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.056445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.056482 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.056494 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.056659 4876 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa" exitCode=0 Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.056704 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa"} Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.056768 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.057733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.057765 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.057775 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.059191 4876 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0" exitCode=0 Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.059411 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0"} Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.059627 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.060956 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.061015 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.061035 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.063849 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03"} Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.063876 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1"} Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.063887 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2"} Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.063898 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e"} Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.063970 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.064863 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.064885 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.064893 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.070611 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a" exitCode=0 Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.070698 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a"} Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.070819 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.072078 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.072132 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.072150 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.077297 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.079762 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.079813 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.079825 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:03 crc kubenswrapper[4876]: E0313 07:39:03.447630 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.39:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189c56929cc9f0be default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,LastTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:03 crc kubenswrapper[4876]: I0313 07:39:03.989546 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:03 crc kubenswrapper[4876]: E0313 07:39:03.993188 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="3.2s" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.077221 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.077280 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.077290 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.077298 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.079272 4876 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e" exitCode=0 Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.079315 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.079415 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.080286 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.080305 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.080314 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.083609 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fa61495c58e2689c63efe08ae1a16e0665a43226d78d3b29d7dd6251b67f8194"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.083680 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.084436 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.084467 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.084474 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.086456 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.086756 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087020 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087040 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087050 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3"} Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087487 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087505 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087512 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087925 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087942 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.087949 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.248250 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.249279 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.249314 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.249324 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:04 crc kubenswrapper[4876]: I0313 07:39:04.249352 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:04 crc kubenswrapper[4876]: E0313 07:39:04.249823 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Mar 13 07:39:04 crc kubenswrapper[4876]: W0313 07:39:04.324650 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:04 crc kubenswrapper[4876]: E0313 07:39:04.324719 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:04 crc kubenswrapper[4876]: W0313 07:39:04.334565 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:39:04 crc kubenswrapper[4876]: E0313 07:39:04.334629 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.094348 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bb9a7c1ad3968cb4926292d3667b21055962ab974c4f2f087a1b3053de840a78"} Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.094513 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.096372 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.096403 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.096421 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.098281 4876 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7" exitCode=0 Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.098503 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.098666 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.099194 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7"} Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.099469 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.099509 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.100165 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.100205 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.100215 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.100426 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.100489 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.101045 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.101088 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.101147 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.101205 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.656014 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.656224 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.657188 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.657214 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.657222 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:05 crc kubenswrapper[4876]: I0313 07:39:05.667092 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.107347 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.107572 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"21de48bddda9ae4ebb09f62645f9287cf815a3d342162be3a7f27209e485b85a"} Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.107620 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d84ec8a00c010220e0d0f5374f11365221798e1fea38320dd6ef538a46279632"} Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.107641 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d760be87350f9f6de9d09d1164320f34913b248d88551440875a177d8f2d098c"} Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.107658 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3a40f3daa61f0648a69b8302e64631a39a78f48a14bc0cceaf946101162c82c0"} Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.107662 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.107798 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.107675 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c022b915c81f918ec768d48e9698848b094036d5de47c50b25615faf3bfe276a"} Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.108599 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109215 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109282 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109346 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109355 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109412 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109298 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109469 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109432 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.109374 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.228321 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:06 crc kubenswrapper[4876]: I0313 07:39:06.953666 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.110806 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.110901 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.110962 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.111062 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.112929 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.112989 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.113009 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.113175 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.113219 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.113273 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.113309 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.113348 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.113364 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.449983 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.451909 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.451993 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.452027 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.452072 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.773916 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:07 crc kubenswrapper[4876]: I0313 07:39:07.994703 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:08 crc kubenswrapper[4876]: I0313 07:39:08.113173 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:08 crc kubenswrapper[4876]: I0313 07:39:08.114117 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:08 crc kubenswrapper[4876]: I0313 07:39:08.114161 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:08 crc kubenswrapper[4876]: I0313 07:39:08.114176 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.024174 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.024363 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.024417 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.025602 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.025652 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.025670 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.115689 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.116982 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.117046 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.117055 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.233103 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.233389 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.234749 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.234779 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:09 crc kubenswrapper[4876]: I0313 07:39:09.234788 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.004967 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.118746 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.120009 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.120067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.120086 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.212442 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.212703 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.214409 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.214491 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:10 crc kubenswrapper[4876]: I0313 07:39:10.214529 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:11 crc kubenswrapper[4876]: I0313 07:39:11.114425 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:11 crc kubenswrapper[4876]: I0313 07:39:11.114604 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:11 crc kubenswrapper[4876]: I0313 07:39:11.115835 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:11 crc kubenswrapper[4876]: I0313 07:39:11.115887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:11 crc kubenswrapper[4876]: I0313 07:39:11.115904 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:11 crc kubenswrapper[4876]: E0313 07:39:11.145854 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:39:12 crc kubenswrapper[4876]: I0313 07:39:12.025314 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:39:12 crc kubenswrapper[4876]: I0313 07:39:12.025411 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 07:39:14 crc kubenswrapper[4876]: I0313 07:39:14.990058 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 13 07:39:15 crc kubenswrapper[4876]: W0313 07:39:15.000338 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.000420 4876 trace.go:236] Trace[570690398]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Mar-2026 07:39:04.998) (total time: 10001ms): Mar 13 07:39:15 crc kubenswrapper[4876]: Trace[570690398]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:39:15.000) Mar 13 07:39:15 crc kubenswrapper[4876]: Trace[570690398]: [10.00152986s] [10.00152986s] END Mar 13 07:39:15 crc kubenswrapper[4876]: E0313 07:39:15.000438 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.133601 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.135447 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bb9a7c1ad3968cb4926292d3667b21055962ab974c4f2f087a1b3053de840a78" exitCode=255 Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.135502 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"bb9a7c1ad3968cb4926292d3667b21055962ab974c4f2f087a1b3053de840a78"} Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.135678 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.136757 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.136798 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.136810 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.137445 4876 scope.go:117] "RemoveContainer" containerID="bb9a7c1ad3968cb4926292d3667b21055962ab974c4f2f087a1b3053de840a78" Mar 13 07:39:15 crc kubenswrapper[4876]: W0313 07:39:15.290997 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.291081 4876 trace.go:236] Trace[841418064]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Mar-2026 07:39:05.289) (total time: 10001ms): Mar 13 07:39:15 crc kubenswrapper[4876]: Trace[841418064]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:39:15.290) Mar 13 07:39:15 crc kubenswrapper[4876]: Trace[841418064]: [10.001834063s] [10.001834063s] END Mar 13 07:39:15 crc kubenswrapper[4876]: E0313 07:39:15.291102 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 13 07:39:15 crc kubenswrapper[4876]: E0313 07:39:15.597624 4876 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:15 crc kubenswrapper[4876]: E0313 07:39:15.598339 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 13 07:39:15 crc kubenswrapper[4876]: W0313 07:39:15.598595 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z Mar 13 07:39:15 crc kubenswrapper[4876]: E0313 07:39:15.598645 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:15 crc kubenswrapper[4876]: E0313 07:39:15.601872 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 07:39:15 crc kubenswrapper[4876]: E0313 07:39:15.603899 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c56929cc9f0be default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,LastTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:15 crc kubenswrapper[4876]: W0313 07:39:15.605946 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z Mar 13 07:39:15 crc kubenswrapper[4876]: E0313 07:39:15.606003 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.606624 4876 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.606690 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.610753 4876 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.610803 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 07:39:15 crc kubenswrapper[4876]: I0313 07:39:15.991702 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:15Z is after 2026-02-23T05:33:13Z Mar 13 07:39:16 crc kubenswrapper[4876]: I0313 07:39:16.140203 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 07:39:16 crc kubenswrapper[4876]: I0313 07:39:16.142089 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b"} Mar 13 07:39:16 crc kubenswrapper[4876]: I0313 07:39:16.142267 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:16 crc kubenswrapper[4876]: I0313 07:39:16.143110 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:16 crc kubenswrapper[4876]: I0313 07:39:16.143137 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:16 crc kubenswrapper[4876]: I0313 07:39:16.143148 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:16 crc kubenswrapper[4876]: I0313 07:39:16.991867 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:16Z is after 2026-02-23T05:33:13Z Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.151266 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.151866 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.157072 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b" exitCode=255 Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.157129 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b"} Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.157202 4876 scope.go:117] "RemoveContainer" containerID="bb9a7c1ad3968cb4926292d3667b21055962ab974c4f2f087a1b3053de840a78" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.157377 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.158289 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.158317 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.158327 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.158746 4876 scope.go:117] "RemoveContainer" containerID="6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b" Mar 13 07:39:17 crc kubenswrapper[4876]: E0313 07:39:17.158899 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.780458 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.993200 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:17Z is after 2026-02-23T05:33:13Z Mar 13 07:39:17 crc kubenswrapper[4876]: I0313 07:39:17.998409 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:18 crc kubenswrapper[4876]: I0313 07:39:18.161626 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 07:39:18 crc kubenswrapper[4876]: I0313 07:39:18.163777 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:18 crc kubenswrapper[4876]: I0313 07:39:18.164517 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:18 crc kubenswrapper[4876]: I0313 07:39:18.164556 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:18 crc kubenswrapper[4876]: I0313 07:39:18.164571 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:18 crc kubenswrapper[4876]: I0313 07:39:18.164984 4876 scope.go:117] "RemoveContainer" containerID="6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b" Mar 13 07:39:18 crc kubenswrapper[4876]: E0313 07:39:18.165158 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:18 crc kubenswrapper[4876]: I0313 07:39:18.992735 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:18Z is after 2026-02-23T05:33:13Z Mar 13 07:39:19 crc kubenswrapper[4876]: I0313 07:39:19.166911 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:19 crc kubenswrapper[4876]: I0313 07:39:19.168268 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:19 crc kubenswrapper[4876]: I0313 07:39:19.168336 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:19 crc kubenswrapper[4876]: I0313 07:39:19.168360 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:19 crc kubenswrapper[4876]: I0313 07:39:19.169227 4876 scope.go:117] "RemoveContainer" containerID="6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b" Mar 13 07:39:19 crc kubenswrapper[4876]: E0313 07:39:19.169605 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:19 crc kubenswrapper[4876]: I0313 07:39:19.994700 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:19Z is after 2026-02-23T05:33:13Z Mar 13 07:39:20 crc kubenswrapper[4876]: W0313 07:39:20.019500 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:20Z is after 2026-02-23T05:33:13Z Mar 13 07:39:20 crc kubenswrapper[4876]: E0313 07:39:20.019616 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:20Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.042315 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.042567 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.044204 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.044329 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.044370 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.063170 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.169619 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.171606 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.171645 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.171657 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:20 crc kubenswrapper[4876]: W0313 07:39:20.178918 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:20Z is after 2026-02-23T05:33:13Z Mar 13 07:39:20 crc kubenswrapper[4876]: E0313 07:39:20.179025 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:20Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:20 crc kubenswrapper[4876]: I0313 07:39:20.991810 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:20Z is after 2026-02-23T05:33:13Z Mar 13 07:39:21 crc kubenswrapper[4876]: I0313 07:39:21.118687 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:21 crc kubenswrapper[4876]: I0313 07:39:21.118812 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:21 crc kubenswrapper[4876]: I0313 07:39:21.119950 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:21 crc kubenswrapper[4876]: I0313 07:39:21.119984 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:21 crc kubenswrapper[4876]: I0313 07:39:21.119993 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:21 crc kubenswrapper[4876]: E0313 07:39:21.146322 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:39:21 crc kubenswrapper[4876]: I0313 07:39:21.991810 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:21Z is after 2026-02-23T05:33:13Z Mar 13 07:39:22 crc kubenswrapper[4876]: E0313 07:39:22.001377 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:22Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 07:39:22 crc kubenswrapper[4876]: I0313 07:39:22.002410 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:22 crc kubenswrapper[4876]: I0313 07:39:22.003468 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:22 crc kubenswrapper[4876]: I0313 07:39:22.003503 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:22 crc kubenswrapper[4876]: I0313 07:39:22.003514 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:22 crc kubenswrapper[4876]: I0313 07:39:22.003535 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:22 crc kubenswrapper[4876]: E0313 07:39:22.005806 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:22Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 07:39:22 crc kubenswrapper[4876]: I0313 07:39:22.025999 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:39:22 crc kubenswrapper[4876]: I0313 07:39:22.026135 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 07:39:22 crc kubenswrapper[4876]: I0313 07:39:22.994350 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:22Z is after 2026-02-23T05:33:13Z Mar 13 07:39:23 crc kubenswrapper[4876]: I0313 07:39:23.442127 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:23 crc kubenswrapper[4876]: I0313 07:39:23.442376 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:23 crc kubenswrapper[4876]: I0313 07:39:23.443621 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:23 crc kubenswrapper[4876]: I0313 07:39:23.443684 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:23 crc kubenswrapper[4876]: I0313 07:39:23.443700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:23 crc kubenswrapper[4876]: I0313 07:39:23.444508 4876 scope.go:117] "RemoveContainer" containerID="6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b" Mar 13 07:39:23 crc kubenswrapper[4876]: E0313 07:39:23.444777 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:23 crc kubenswrapper[4876]: I0313 07:39:23.546567 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:23 crc kubenswrapper[4876]: I0313 07:39:23.994091 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:23Z is after 2026-02-23T05:33:13Z Mar 13 07:39:24 crc kubenswrapper[4876]: I0313 07:39:24.179854 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:24 crc kubenswrapper[4876]: I0313 07:39:24.181137 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:24 crc kubenswrapper[4876]: I0313 07:39:24.181166 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:24 crc kubenswrapper[4876]: I0313 07:39:24.181178 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:24 crc kubenswrapper[4876]: I0313 07:39:24.181745 4876 scope.go:117] "RemoveContainer" containerID="6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b" Mar 13 07:39:24 crc kubenswrapper[4876]: E0313 07:39:24.182003 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:24 crc kubenswrapper[4876]: I0313 07:39:24.289704 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 07:39:24 crc kubenswrapper[4876]: E0313 07:39:24.293572 4876 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:24 crc kubenswrapper[4876]: I0313 07:39:24.993858 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:24Z is after 2026-02-23T05:33:13Z Mar 13 07:39:25 crc kubenswrapper[4876]: W0313 07:39:25.140873 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:25Z is after 2026-02-23T05:33:13Z Mar 13 07:39:25 crc kubenswrapper[4876]: E0313 07:39:25.141358 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:25Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:25 crc kubenswrapper[4876]: E0313 07:39:25.607410 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:25Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c56929cc9f0be default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,LastTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:25 crc kubenswrapper[4876]: I0313 07:39:25.993692 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:25Z is after 2026-02-23T05:33:13Z Mar 13 07:39:26 crc kubenswrapper[4876]: W0313 07:39:26.469421 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:26Z is after 2026-02-23T05:33:13Z Mar 13 07:39:26 crc kubenswrapper[4876]: E0313 07:39:26.469565 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:26 crc kubenswrapper[4876]: W0313 07:39:26.978372 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:26Z is after 2026-02-23T05:33:13Z Mar 13 07:39:26 crc kubenswrapper[4876]: E0313 07:39:26.978691 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:26 crc kubenswrapper[4876]: I0313 07:39:26.991820 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:26Z is after 2026-02-23T05:33:13Z Mar 13 07:39:27 crc kubenswrapper[4876]: W0313 07:39:27.554029 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:27Z is after 2026-02-23T05:33:13Z Mar 13 07:39:27 crc kubenswrapper[4876]: E0313 07:39:27.554115 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:27Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:27 crc kubenswrapper[4876]: I0313 07:39:27.992020 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:27Z is after 2026-02-23T05:33:13Z Mar 13 07:39:28 crc kubenswrapper[4876]: I0313 07:39:28.991801 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:28Z is after 2026-02-23T05:33:13Z Mar 13 07:39:29 crc kubenswrapper[4876]: E0313 07:39:29.005676 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:29Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 07:39:29 crc kubenswrapper[4876]: I0313 07:39:29.006749 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:29 crc kubenswrapper[4876]: I0313 07:39:29.007674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:29 crc kubenswrapper[4876]: I0313 07:39:29.007708 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:29 crc kubenswrapper[4876]: I0313 07:39:29.007719 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:29 crc kubenswrapper[4876]: I0313 07:39:29.007740 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:29 crc kubenswrapper[4876]: E0313 07:39:29.010316 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:29Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 07:39:29 crc kubenswrapper[4876]: I0313 07:39:29.992121 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:29Z is after 2026-02-23T05:33:13Z Mar 13 07:39:30 crc kubenswrapper[4876]: I0313 07:39:30.992161 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:30Z is after 2026-02-23T05:33:13Z Mar 13 07:39:31 crc kubenswrapper[4876]: E0313 07:39:31.147071 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:39:31 crc kubenswrapper[4876]: I0313 07:39:31.992135 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.025369 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.025462 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.025541 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.025708 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.029108 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.029342 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.029379 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.031089 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.031301 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2" gracePeriod=30 Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.200890 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.201318 4876 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2" exitCode=255 Mar 13 07:39:32 crc kubenswrapper[4876]: I0313 07:39:32.201358 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2"} Mar 13 07:39:33 crc kubenswrapper[4876]: I0313 07:39:33.108407 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:33Z is after 2026-02-23T05:33:13Z Mar 13 07:39:33 crc kubenswrapper[4876]: I0313 07:39:33.208505 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 07:39:33 crc kubenswrapper[4876]: I0313 07:39:33.209028 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146"} Mar 13 07:39:33 crc kubenswrapper[4876]: I0313 07:39:33.209170 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:33 crc kubenswrapper[4876]: I0313 07:39:33.210370 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:33 crc kubenswrapper[4876]: I0313 07:39:33.210435 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:33 crc kubenswrapper[4876]: I0313 07:39:33.210448 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:33 crc kubenswrapper[4876]: I0313 07:39:33.993021 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:33Z is after 2026-02-23T05:33:13Z Mar 13 07:39:34 crc kubenswrapper[4876]: I0313 07:39:34.211142 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:34 crc kubenswrapper[4876]: I0313 07:39:34.212019 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:34 crc kubenswrapper[4876]: I0313 07:39:34.212069 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:34 crc kubenswrapper[4876]: I0313 07:39:34.212084 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:34 crc kubenswrapper[4876]: I0313 07:39:34.992083 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:34Z is after 2026-02-23T05:33:13Z Mar 13 07:39:35 crc kubenswrapper[4876]: E0313 07:39:35.612730 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:35Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c56929cc9f0be default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,LastTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:35 crc kubenswrapper[4876]: I0313 07:39:35.991903 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:35Z is after 2026-02-23T05:33:13Z Mar 13 07:39:36 crc kubenswrapper[4876]: E0313 07:39:36.009624 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:36Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.010824 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.012260 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.012304 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.012316 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.012340 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:36 crc kubenswrapper[4876]: E0313 07:39:36.016697 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:36Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.034955 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.036070 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.036111 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.036123 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.036665 4876 scope.go:117] "RemoveContainer" containerID="6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.218166 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.219795 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186"} Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.220008 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.220861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.220895 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.220907 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.229145 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.229323 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.230291 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.230321 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.230330 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:36 crc kubenswrapper[4876]: I0313 07:39:36.993637 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:36Z is after 2026-02-23T05:33:13Z Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.224266 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.224767 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.226456 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186" exitCode=255 Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.226496 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186"} Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.226540 4876 scope.go:117] "RemoveContainer" containerID="6cff69cd98ad3f23cd1ab20a40d80781c60dab7ae011894fa004d562fea7571b" Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.226629 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.227659 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.227683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.227693 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.228212 4876 scope.go:117] "RemoveContainer" containerID="9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186" Mar 13 07:39:37 crc kubenswrapper[4876]: E0313 07:39:37.228399 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:37 crc kubenswrapper[4876]: I0313 07:39:37.992808 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:37Z is after 2026-02-23T05:33:13Z Mar 13 07:39:38 crc kubenswrapper[4876]: I0313 07:39:38.230182 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 07:39:38 crc kubenswrapper[4876]: I0313 07:39:38.992651 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:38Z is after 2026-02-23T05:33:13Z Mar 13 07:39:39 crc kubenswrapper[4876]: I0313 07:39:39.025101 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:39 crc kubenswrapper[4876]: I0313 07:39:39.025375 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:39 crc kubenswrapper[4876]: I0313 07:39:39.026722 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:39 crc kubenswrapper[4876]: I0313 07:39:39.026768 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:39 crc kubenswrapper[4876]: I0313 07:39:39.026784 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:39 crc kubenswrapper[4876]: W0313 07:39:39.547186 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:39Z is after 2026-02-23T05:33:13Z Mar 13 07:39:39 crc kubenswrapper[4876]: E0313 07:39:39.547262 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:39Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:39 crc kubenswrapper[4876]: I0313 07:39:39.991396 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:39Z is after 2026-02-23T05:33:13Z Mar 13 07:39:40 crc kubenswrapper[4876]: I0313 07:39:40.558123 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 07:39:40 crc kubenswrapper[4876]: E0313 07:39:40.564130 4876 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:40Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:40 crc kubenswrapper[4876]: E0313 07:39:40.565312 4876 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 13 07:39:40 crc kubenswrapper[4876]: I0313 07:39:40.992405 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:40Z is after 2026-02-23T05:33:13Z Mar 13 07:39:41 crc kubenswrapper[4876]: E0313 07:39:41.147278 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:39:41 crc kubenswrapper[4876]: I0313 07:39:41.991768 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:41Z is after 2026-02-23T05:33:13Z Mar 13 07:39:42 crc kubenswrapper[4876]: I0313 07:39:42.026158 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:39:42 crc kubenswrapper[4876]: I0313 07:39:42.026268 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 07:39:42 crc kubenswrapper[4876]: I0313 07:39:42.994804 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:42Z is after 2026-02-23T05:33:13Z Mar 13 07:39:43 crc kubenswrapper[4876]: E0313 07:39:43.014898 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:43Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.017149 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.019046 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.019103 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.019121 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.019166 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:43 crc kubenswrapper[4876]: E0313 07:39:43.022130 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:43Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.442030 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.442214 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.443333 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.443383 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.443394 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.443869 4876 scope.go:117] "RemoveContainer" containerID="9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186" Mar 13 07:39:43 crc kubenswrapper[4876]: E0313 07:39:43.444021 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.546631 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:39:43 crc kubenswrapper[4876]: I0313 07:39:43.991653 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:43Z is after 2026-02-23T05:33:13Z Mar 13 07:39:44 crc kubenswrapper[4876]: I0313 07:39:44.246064 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:44 crc kubenswrapper[4876]: I0313 07:39:44.246965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:44 crc kubenswrapper[4876]: I0313 07:39:44.247011 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:44 crc kubenswrapper[4876]: I0313 07:39:44.247023 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:44 crc kubenswrapper[4876]: I0313 07:39:44.247583 4876 scope.go:117] "RemoveContainer" containerID="9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186" Mar 13 07:39:44 crc kubenswrapper[4876]: E0313 07:39:44.247811 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:44 crc kubenswrapper[4876]: I0313 07:39:44.994018 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:44Z is after 2026-02-23T05:33:13Z Mar 13 07:39:45 crc kubenswrapper[4876]: W0313 07:39:45.097722 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:45Z is after 2026-02-23T05:33:13Z Mar 13 07:39:45 crc kubenswrapper[4876]: E0313 07:39:45.097824 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:45Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 07:39:45 crc kubenswrapper[4876]: E0313 07:39:45.618058 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:45Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c56929cc9f0be default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,LastTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:45 crc kubenswrapper[4876]: I0313 07:39:45.991704 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:45Z is after 2026-02-23T05:33:13Z Mar 13 07:39:46 crc kubenswrapper[4876]: I0313 07:39:46.993471 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:47 crc kubenswrapper[4876]: W0313 07:39:47.779096 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 13 07:39:47 crc kubenswrapper[4876]: E0313 07:39:47.779174 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 07:39:48 crc kubenswrapper[4876]: I0313 07:39:48.000959 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:48 crc kubenswrapper[4876]: I0313 07:39:48.996477 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:49 crc kubenswrapper[4876]: W0313 07:39:49.170402 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 13 07:39:49 crc kubenswrapper[4876]: E0313 07:39:49.170483 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 07:39:49 crc kubenswrapper[4876]: I0313 07:39:49.995972 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:50 crc kubenswrapper[4876]: E0313 07:39:50.020525 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.022535 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.027199 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.027265 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.027277 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.027309 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:50 crc kubenswrapper[4876]: E0313 07:39:50.033315 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.217376 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.217640 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.219202 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.219293 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.219314 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:50 crc kubenswrapper[4876]: I0313 07:39:50.996653 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:51 crc kubenswrapper[4876]: E0313 07:39:51.147543 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:39:51 crc kubenswrapper[4876]: I0313 07:39:51.996312 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:52 crc kubenswrapper[4876]: I0313 07:39:52.025514 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:39:52 crc kubenswrapper[4876]: I0313 07:39:52.025625 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 07:39:52 crc kubenswrapper[4876]: I0313 07:39:52.994478 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:53 crc kubenswrapper[4876]: I0313 07:39:53.996027 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:55 crc kubenswrapper[4876]: I0313 07:39:55.014742 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.625851 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c56929cc9f0be default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,LastTimestamp:2026-03-13 07:39:00.972826814 +0000 UTC m=+0.643605826,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.631648 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f93917 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,LastTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.640226 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f96fef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,LastTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.646966 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f998fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043058941 +0000 UTC m=+0.713837923,LastTimestamp:2026-03-13 07:39:01.043058941 +0000 UTC m=+0.713837923,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.652051 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a6a984b4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.138474164 +0000 UTC m=+0.809253156,LastTimestamp:2026-03-13 07:39:01.138474164 +0000 UTC m=+0.809253156,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.656935 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f93917\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f93917 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,LastTimestamp:2026-03-13 07:39:01.236910093 +0000 UTC m=+0.907689095,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.661020 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f96fef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f96fef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,LastTimestamp:2026-03-13 07:39:01.236932934 +0000 UTC m=+0.907711926,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.665085 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f998fd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f998fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043058941 +0000 UTC m=+0.713837923,LastTimestamp:2026-03-13 07:39:01.236943844 +0000 UTC m=+0.907722836,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.670226 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f93917\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f93917 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,LastTimestamp:2026-03-13 07:39:01.337214195 +0000 UTC m=+1.007993187,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.675212 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f96fef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f96fef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,LastTimestamp:2026-03-13 07:39:01.337266667 +0000 UTC m=+1.008045669,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.680875 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f998fd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f998fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043058941 +0000 UTC m=+0.713837923,LastTimestamp:2026-03-13 07:39:01.337293548 +0000 UTC m=+1.008072540,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.685614 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f93917\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f93917 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,LastTimestamp:2026-03-13 07:39:01.33864566 +0000 UTC m=+1.009424642,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.690318 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f96fef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f96fef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,LastTimestamp:2026-03-13 07:39:01.338666141 +0000 UTC m=+1.009445113,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.694859 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f998fd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f998fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043058941 +0000 UTC m=+0.713837923,LastTimestamp:2026-03-13 07:39:01.338676281 +0000 UTC m=+1.009455263,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.699030 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f93917\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f93917 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,LastTimestamp:2026-03-13 07:39:01.339446351 +0000 UTC m=+1.010225343,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.703353 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f96fef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f96fef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,LastTimestamp:2026-03-13 07:39:01.339465141 +0000 UTC m=+1.010244133,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.708187 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f998fd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f998fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043058941 +0000 UTC m=+0.713837923,LastTimestamp:2026-03-13 07:39:01.339478212 +0000 UTC m=+1.010257214,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.713356 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f93917\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f93917 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,LastTimestamp:2026-03-13 07:39:01.339870117 +0000 UTC m=+1.010649129,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.717779 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f96fef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f96fef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,LastTimestamp:2026-03-13 07:39:01.339889187 +0000 UTC m=+1.010668179,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.721853 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f998fd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f998fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043058941 +0000 UTC m=+0.713837923,LastTimestamp:2026-03-13 07:39:01.339901878 +0000 UTC m=+1.010680870,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.726407 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f93917\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f93917 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,LastTimestamp:2026-03-13 07:39:01.340163679 +0000 UTC m=+1.010942671,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.731711 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f96fef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f96fef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,LastTimestamp:2026-03-13 07:39:01.340179359 +0000 UTC m=+1.010958362,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.737315 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f998fd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f998fd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043058941 +0000 UTC m=+0.713837923,LastTimestamp:2026-03-13 07:39:01.34019179 +0000 UTC m=+1.010970782,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.741014 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f93917\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f93917 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043034391 +0000 UTC m=+0.713813373,LastTimestamp:2026-03-13 07:39:01.341413627 +0000 UTC m=+1.012192599,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.747498 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5692a0f96fef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5692a0f96fef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.043048431 +0000 UTC m=+0.713827413,LastTimestamp:2026-03-13 07:39:01.34147503 +0000 UTC m=+1.012254012,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.752525 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5692cab793a8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.743375272 +0000 UTC m=+1.414154264,LastTimestamp:2026-03-13 07:39:01.743375272 +0000 UTC m=+1.414154264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.756695 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692cab97e9d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.743500957 +0000 UTC m=+1.414279939,LastTimestamp:2026-03-13 07:39:01.743500957 +0000 UTC m=+1.414279939,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.761573 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5692cb77b8b5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.755967669 +0000 UTC m=+1.426746661,LastTimestamp:2026-03-13 07:39:01.755967669 +0000 UTC m=+1.426746661,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.766220 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5692cc270e78 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.767458424 +0000 UTC m=+1.438237406,LastTimestamp:2026-03-13 07:39:01.767458424 +0000 UTC m=+1.438237406,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.770916 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5692cd321d90 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:01.7849604 +0000 UTC m=+1.455739392,LastTimestamp:2026-03-13 07:39:01.7849604 +0000 UTC m=+1.455739392,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.775104 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5692eaff24c3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.284936387 +0000 UTC m=+1.955715369,LastTimestamp:2026-03-13 07:39:02.284936387 +0000 UTC m=+1.955715369,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.778858 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5692eb002e5c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.28500438 +0000 UTC m=+1.955783362,LastTimestamp:2026-03-13 07:39:02.28500438 +0000 UTC m=+1.955783362,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.782157 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5692eb01403e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.285074494 +0000 UTC m=+1.955853476,LastTimestamp:2026-03-13 07:39:02.285074494 +0000 UTC m=+1.955853476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.786467 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5692eb03a149 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.285230409 +0000 UTC m=+1.956009391,LastTimestamp:2026-03-13 07:39:02.285230409 +0000 UTC m=+1.956009391,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.789973 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692eb07767f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.285481599 +0000 UTC m=+1.956260581,LastTimestamp:2026-03-13 07:39:02.285481599 +0000 UTC m=+1.956260581,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.793717 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5692ebadc2b1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.296380081 +0000 UTC m=+1.967159063,LastTimestamp:2026-03-13 07:39:02.296380081 +0000 UTC m=+1.967159063,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.797303 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5692ebea3f45 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.300344133 +0000 UTC m=+1.971123115,LastTimestamp:2026-03-13 07:39:02.300344133 +0000 UTC m=+1.971123115,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.801195 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5692ec1bfdb6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.30360415 +0000 UTC m=+1.974383132,LastTimestamp:2026-03-13 07:39:02.30360415 +0000 UTC m=+1.974383132,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.805287 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692ec203657 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.303880791 +0000 UTC m=+1.974659773,LastTimestamp:2026-03-13 07:39:02.303880791 +0000 UTC m=+1.974659773,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.808810 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5692ec270de4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.304329188 +0000 UTC m=+1.975108170,LastTimestamp:2026-03-13 07:39:02.304329188 +0000 UTC m=+1.975108170,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.812692 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692ec3bc1d2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.30568597 +0000 UTC m=+1.976464952,LastTimestamp:2026-03-13 07:39:02.30568597 +0000 UTC m=+1.976464952,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.816257 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692fd5b8ea0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.592982688 +0000 UTC m=+2.263761670,LastTimestamp:2026-03-13 07:39:02.592982688 +0000 UTC m=+2.263761670,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.820627 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692fe029535 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.603928885 +0000 UTC m=+2.274707897,LastTimestamp:2026-03-13 07:39:02.603928885 +0000 UTC m=+2.274707897,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.825331 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692fe12e4ca openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.604997834 +0000 UTC m=+2.275776846,LastTimestamp:2026-03-13 07:39:02.604997834 +0000 UTC m=+2.275776846,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.829441 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c569309ef1a14 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.803978772 +0000 UTC m=+2.474757794,LastTimestamp:2026-03-13 07:39:02.803978772 +0000 UTC m=+2.474757794,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.832990 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c56930a87104f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.813937743 +0000 UTC m=+2.484716725,LastTimestamp:2026-03-13 07:39:02.813937743 +0000 UTC m=+2.484716725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.837265 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c56930a987a49 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.815078985 +0000 UTC m=+2.485857967,LastTimestamp:2026-03-13 07:39:02.815078985 +0000 UTC m=+2.485857967,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.840797 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5693145d64d8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.978979032 +0000 UTC m=+2.649758014,LastTimestamp:2026-03-13 07:39:02.978979032 +0000 UTC m=+2.649758014,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.844701 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5693151c37cb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.991484875 +0000 UTC m=+2.662263857,LastTimestamp:2026-03-13 07:39:02.991484875 +0000 UTC m=+2.662263857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.849347 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c569319388d8f openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.060450703 +0000 UTC m=+2.731229685,LastTimestamp:2026-03-13 07:39:03.060450703 +0000 UTC m=+2.731229685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.853391 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c56931939dcec openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.060536556 +0000 UTC m=+2.731315548,LastTimestamp:2026-03-13 07:39:03.060536556 +0000 UTC m=+2.731315548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.861864 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c569319656ad2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.06339093 +0000 UTC m=+2.734169912,LastTimestamp:2026-03-13 07:39:03.06339093 +0000 UTC m=+2.734169912,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.866456 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56931a354cb0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.077014704 +0000 UTC m=+2.747793716,LastTimestamp:2026-03-13 07:39:03.077014704 +0000 UTC m=+2.747793716,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.870784 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5693263706a9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.278454441 +0000 UTC m=+2.949233423,LastTimestamp:2026-03-13 07:39:03.278454441 +0000 UTC m=+2.949233423,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.874676 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c569326690ed2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.28173333 +0000 UTC m=+2.952512312,LastTimestamp:2026-03-13 07:39:03.28173333 +0000 UTC m=+2.952512312,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.878482 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c569326c5083f openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.287760959 +0000 UTC m=+2.958539941,LastTimestamp:2026-03-13 07:39:03.287760959 +0000 UTC m=+2.958539941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.882218 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c569326d621a1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.288881569 +0000 UTC m=+2.959660551,LastTimestamp:2026-03-13 07:39:03.288881569 +0000 UTC m=+2.959660551,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.886625 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c569326e424ca openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.289799882 +0000 UTC m=+2.960578864,LastTimestamp:2026-03-13 07:39:03.289799882 +0000 UTC m=+2.960578864,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.890695 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c569327cef174 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.3051877 +0000 UTC m=+2.975966682,LastTimestamp:2026-03-13 07:39:03.3051877 +0000 UTC m=+2.975966682,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.894057 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5693280f61e9 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.309410793 +0000 UTC m=+2.980189775,LastTimestamp:2026-03-13 07:39:03.309410793 +0000 UTC m=+2.980189775,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.898182 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5693285c2639 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.314441785 +0000 UTC m=+2.985220767,LastTimestamp:2026-03-13 07:39:03.314441785 +0000 UTC m=+2.985220767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.902414 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56932943862d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.329605165 +0000 UTC m=+3.000384157,LastTimestamp:2026-03-13 07:39:03.329605165 +0000 UTC m=+3.000384157,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.905992 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5693295405e1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.330686433 +0000 UTC m=+3.001465415,LastTimestamp:2026-03-13 07:39:03.330686433 +0000 UTC m=+3.001465415,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.909136 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c569332a4e4d8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.486981336 +0000 UTC m=+3.157760318,LastTimestamp:2026-03-13 07:39:03.486981336 +0000 UTC m=+3.157760318,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.912733 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5693337428b2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.500564658 +0000 UTC m=+3.171343640,LastTimestamp:2026-03-13 07:39:03.500564658 +0000 UTC m=+3.171343640,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.916541 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c56933387bda7 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.501847975 +0000 UTC m=+3.172626967,LastTimestamp:2026-03-13 07:39:03.501847975 +0000 UTC m=+3.172626967,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.919767 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c569335a6cb3f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.537437503 +0000 UTC m=+3.208216485,LastTimestamp:2026-03-13 07:39:03.537437503 +0000 UTC m=+3.208216485,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.922930 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56933690d055 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.552774229 +0000 UTC m=+3.223553211,LastTimestamp:2026-03-13 07:39:03.552774229 +0000 UTC m=+3.223553211,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.926105 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c569336a289af openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.553935791 +0000 UTC m=+3.224714783,LastTimestamp:2026-03-13 07:39:03.553935791 +0000 UTC m=+3.224714783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.929358 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c56933e8501b3 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.686218163 +0000 UTC m=+3.356997145,LastTimestamp:2026-03-13 07:39:03.686218163 +0000 UTC m=+3.356997145,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.932702 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c56933f3920fb openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.698022651 +0000 UTC m=+3.368801633,LastTimestamp:2026-03-13 07:39:03.698022651 +0000 UTC m=+3.368801633,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.936112 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c569341e8aab7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.743081143 +0000 UTC m=+3.413860125,LastTimestamp:2026-03-13 07:39:03.743081143 +0000 UTC m=+3.413860125,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.939297 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c569342f8c9ca openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.76091489 +0000 UTC m=+3.431693892,LastTimestamp:2026-03-13 07:39:03.76091489 +0000 UTC m=+3.431693892,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.943490 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5693430e7a41 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.762336321 +0000 UTC m=+3.433115293,LastTimestamp:2026-03-13 07:39:03.762336321 +0000 UTC m=+3.433115293,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.946787 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56934d398ae6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.93293079 +0000 UTC m=+3.603709772,LastTimestamp:2026-03-13 07:39:03.93293079 +0000 UTC m=+3.603709772,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.950620 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56934de7a4fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.944340734 +0000 UTC m=+3.615119716,LastTimestamp:2026-03-13 07:39:03.944340734 +0000 UTC m=+3.615119716,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.953976 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56934df977ee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.945508846 +0000 UTC m=+3.616287838,LastTimestamp:2026-03-13 07:39:03.945508846 +0000 UTC m=+3.616287838,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.957829 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c56935623897a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:04.082483578 +0000 UTC m=+3.753262560,LastTimestamp:2026-03-13 07:39:04.082483578 +0000 UTC m=+3.753262560,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.963213 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56935a99ca71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:04.157342321 +0000 UTC m=+3.828121303,LastTimestamp:2026-03-13 07:39:04.157342321 +0000 UTC m=+3.828121303,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.967181 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56935b7f909c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:04.172400796 +0000 UTC m=+3.843179788,LastTimestamp:2026-03-13 07:39:04.172400796 +0000 UTC m=+3.843179788,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.971661 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c56936319de0d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:04.299953677 +0000 UTC m=+3.970732659,LastTimestamp:2026-03-13 07:39:04.299953677 +0000 UTC m=+3.970732659,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.975493 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c56936429a0c6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:04.317763782 +0000 UTC m=+3.988542774,LastTimestamp:2026-03-13 07:39:04.317763782 +0000 UTC m=+3.988542774,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.983757 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c569392fee48c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.103492236 +0000 UTC m=+4.774271228,LastTimestamp:2026-03-13 07:39:05.103492236 +0000 UTC m=+4.774271228,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.989747 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c56939c7663b8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.26231852 +0000 UTC m=+4.933097502,LastTimestamp:2026-03-13 07:39:05.26231852 +0000 UTC m=+4.933097502,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:55 crc kubenswrapper[4876]: I0313 07:39:55.990026 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:55 crc kubenswrapper[4876]: E0313 07:39:55.996444 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c56939d0c7188 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.272152456 +0000 UTC m=+4.942931458,LastTimestamp:2026-03-13 07:39:05.272152456 +0000 UTC m=+4.942931458,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.000939 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c56939d23157c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.27363622 +0000 UTC m=+4.944415202,LastTimestamp:2026-03-13 07:39:05.27363622 +0000 UTC m=+4.944415202,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.008090 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693a8be8777 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.468372855 +0000 UTC m=+5.139151837,LastTimestamp:2026-03-13 07:39:05.468372855 +0000 UTC m=+5.139151837,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.015054 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693a94512d2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.477190354 +0000 UTC m=+5.147969336,LastTimestamp:2026-03-13 07:39:05.477190354 +0000 UTC m=+5.147969336,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.019602 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693a955f7fa openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.478297594 +0000 UTC m=+5.149076576,LastTimestamp:2026-03-13 07:39:05.478297594 +0000 UTC m=+5.149076576,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.023791 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693b45fe4a7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.663497383 +0000 UTC m=+5.334276365,LastTimestamp:2026-03-13 07:39:05.663497383 +0000 UTC m=+5.334276365,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.026895 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693b51edc7a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.676012666 +0000 UTC m=+5.346791648,LastTimestamp:2026-03-13 07:39:05.676012666 +0000 UTC m=+5.346791648,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.033289 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693b52dd6cc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.676994252 +0000 UTC m=+5.347773224,LastTimestamp:2026-03-13 07:39:05.676994252 +0000 UTC m=+5.347773224,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: I0313 07:39:56.035296 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:56 crc kubenswrapper[4876]: I0313 07:39:56.037156 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:56 crc kubenswrapper[4876]: I0313 07:39:56.037211 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:56 crc kubenswrapper[4876]: I0313 07:39:56.037230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:56 crc kubenswrapper[4876]: I0313 07:39:56.038148 4876 scope.go:117] "RemoveContainer" containerID="9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.038526 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.038978 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693c124b614 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.877722644 +0000 UTC m=+5.548501626,LastTimestamp:2026-03-13 07:39:05.877722644 +0000 UTC m=+5.548501626,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.043722 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693c2088471 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.892652145 +0000 UTC m=+5.563431167,LastTimestamp:2026-03-13 07:39:05.892652145 +0000 UTC m=+5.563431167,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.051273 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693c21a544b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:05.893819467 +0000 UTC m=+5.564598449,LastTimestamp:2026-03-13 07:39:05.893819467 +0000 UTC m=+5.564598449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.058523 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693ccae32d2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:06.071282386 +0000 UTC m=+5.742061388,LastTimestamp:2026-03-13 07:39:06.071282386 +0000 UTC m=+5.742061388,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.065316 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5693cdcaaf98 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:06.089926552 +0000 UTC m=+5.760705564,LastTimestamp:2026-03-13 07:39:06.089926552 +0000 UTC m=+5.760705564,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.071927 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 07:39:56 crc kubenswrapper[4876]: &Event{ObjectMeta:{kube-controller-manager-crc.189c56952f92923d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 13 07:39:56 crc kubenswrapper[4876]: body: Mar 13 07:39:56 crc kubenswrapper[4876]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:12.025383485 +0000 UTC m=+11.696162487,LastTimestamp:2026-03-13 07:39:12.025383485 +0000 UTC m=+11.696162487,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 07:39:56 crc kubenswrapper[4876]: > Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.076869 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c56952f938766 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:12.025446246 +0000 UTC m=+11.696225248,LastTimestamp:2026-03-13 07:39:12.025446246 +0000 UTC m=+11.696225248,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.081369 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c56934df977ee\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56934df977ee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:03.945508846 +0000 UTC m=+3.616287838,LastTimestamp:2026-03-13 07:39:15.138450506 +0000 UTC m=+14.809229488,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.087103 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c56935a99ca71\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56935a99ca71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:04.157342321 +0000 UTC m=+3.828121303,LastTimestamp:2026-03-13 07:39:15.293326149 +0000 UTC m=+14.964105131,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.091921 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c56935b7f909c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c56935b7f909c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:04.172400796 +0000 UTC m=+3.843179788,LastTimestamp:2026-03-13 07:39:15.313987647 +0000 UTC m=+14.984766629,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.097968 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 07:39:56 crc kubenswrapper[4876]: &Event{ObjectMeta:{kube-apiserver-crc.189c56960508b52b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 13 07:39:56 crc kubenswrapper[4876]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 07:39:56 crc kubenswrapper[4876]: Mar 13 07:39:56 crc kubenswrapper[4876]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:15.606672683 +0000 UTC m=+15.277451665,LastTimestamp:2026-03-13 07:39:15.606672683 +0000 UTC m=+15.277451665,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 07:39:56 crc kubenswrapper[4876]: > Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.103459 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5696050958ec openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:15.606714604 +0000 UTC m=+15.277493586,LastTimestamp:2026-03-13 07:39:15.606714604 +0000 UTC m=+15.277493586,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.108315 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c56960508b52b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 07:39:56 crc kubenswrapper[4876]: &Event{ObjectMeta:{kube-apiserver-crc.189c56960508b52b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 13 07:39:56 crc kubenswrapper[4876]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 07:39:56 crc kubenswrapper[4876]: Mar 13 07:39:56 crc kubenswrapper[4876]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:15.606672683 +0000 UTC m=+15.277451665,LastTimestamp:2026-03-13 07:39:15.610790004 +0000 UTC m=+15.281568986,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 07:39:56 crc kubenswrapper[4876]: > Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.114145 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5696050958ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5696050958ec openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:15.606714604 +0000 UTC m=+15.277493586,LastTimestamp:2026-03-13 07:39:15.610822535 +0000 UTC m=+15.281601517,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.121335 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 07:39:56 crc kubenswrapper[4876]: &Event{ObjectMeta:{kube-controller-manager-crc.189c569783a990bd openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 07:39:56 crc kubenswrapper[4876]: body: Mar 13 07:39:56 crc kubenswrapper[4876]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:22.026111165 +0000 UTC m=+21.696890167,LastTimestamp:2026-03-13 07:39:22.026111165 +0000 UTC m=+21.696890167,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 07:39:56 crc kubenswrapper[4876]: > Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.125685 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c569783aa5301 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:22.026160897 +0000 UTC m=+21.696939879,LastTimestamp:2026-03-13 07:39:22.026160897 +0000 UTC m=+21.696939879,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.131718 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c569783a990bd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 07:39:56 crc kubenswrapper[4876]: &Event{ObjectMeta:{kube-controller-manager-crc.189c569783a990bd openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 07:39:56 crc kubenswrapper[4876]: body: Mar 13 07:39:56 crc kubenswrapper[4876]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:22.026111165 +0000 UTC m=+21.696890167,LastTimestamp:2026-03-13 07:39:32.025429514 +0000 UTC m=+31.696208496,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 07:39:56 crc kubenswrapper[4876]: > Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.136377 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c569783aa5301\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c569783aa5301 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:22.026160897 +0000 UTC m=+21.696939879,LastTimestamp:2026-03-13 07:39:32.025506036 +0000 UTC m=+31.696285018,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.143990 4876 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5699d8044717 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:32.031276823 +0000 UTC m=+31.702059325,LastTimestamp:2026-03-13 07:39:32.031276823 +0000 UTC m=+31.702059325,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.148767 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5692ec3bc1d2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692ec3bc1d2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.30568597 +0000 UTC m=+1.976464952,LastTimestamp:2026-03-13 07:39:32.148672943 +0000 UTC m=+31.819451925,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.156538 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5692fd5b8ea0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692fd5b8ea0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.592982688 +0000 UTC m=+2.263761670,LastTimestamp:2026-03-13 07:39:32.336462567 +0000 UTC m=+32.007241559,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.161104 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5692fe029535\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5692fe029535 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:02.603928885 +0000 UTC m=+2.274707897,LastTimestamp:2026-03-13 07:39:32.349340714 +0000 UTC m=+32.020119696,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.170558 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c569783a990bd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 07:39:56 crc kubenswrapper[4876]: &Event{ObjectMeta:{kube-controller-manager-crc.189c569783a990bd openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 07:39:56 crc kubenswrapper[4876]: body: Mar 13 07:39:56 crc kubenswrapper[4876]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:22.026111165 +0000 UTC m=+21.696890167,LastTimestamp:2026-03-13 07:39:42.026218646 +0000 UTC m=+41.696997638,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 07:39:56 crc kubenswrapper[4876]: > Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.184404 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c569783aa5301\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c569783aa5301 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:22.026160897 +0000 UTC m=+21.696939879,LastTimestamp:2026-03-13 07:39:42.026304409 +0000 UTC m=+41.697083411,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:39:56 crc kubenswrapper[4876]: E0313 07:39:56.193609 4876 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c569783a990bd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 07:39:56 crc kubenswrapper[4876]: &Event{ObjectMeta:{kube-controller-manager-crc.189c569783a990bd openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 07:39:56 crc kubenswrapper[4876]: body: Mar 13 07:39:56 crc kubenswrapper[4876]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:39:22.026111165 +0000 UTC m=+21.696890167,LastTimestamp:2026-03-13 07:39:52.02558588 +0000 UTC m=+51.696364932,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 07:39:56 crc kubenswrapper[4876]: > Mar 13 07:39:56 crc kubenswrapper[4876]: I0313 07:39:56.993430 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:57 crc kubenswrapper[4876]: E0313 07:39:57.025205 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 07:39:57 crc kubenswrapper[4876]: I0313 07:39:57.034318 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:57 crc kubenswrapper[4876]: I0313 07:39:57.035307 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:57 crc kubenswrapper[4876]: I0313 07:39:57.035341 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:57 crc kubenswrapper[4876]: I0313 07:39:57.035352 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:57 crc kubenswrapper[4876]: I0313 07:39:57.035374 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:39:57 crc kubenswrapper[4876]: E0313 07:39:57.038927 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 07:39:57 crc kubenswrapper[4876]: I0313 07:39:57.994538 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:58 crc kubenswrapper[4876]: I0313 07:39:58.995996 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.032016 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.032173 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.033634 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.033674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.033686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.038716 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.290872 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.292309 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.292381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.292394 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:39:59 crc kubenswrapper[4876]: I0313 07:39:59.993547 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:00 crc kubenswrapper[4876]: I0313 07:40:00.993699 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:01 crc kubenswrapper[4876]: E0313 07:40:01.147782 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:40:01 crc kubenswrapper[4876]: I0313 07:40:01.992990 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:02 crc kubenswrapper[4876]: I0313 07:40:02.994531 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:03 crc kubenswrapper[4876]: I0313 07:40:03.992311 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:04 crc kubenswrapper[4876]: E0313 07:40:04.029544 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 07:40:04 crc kubenswrapper[4876]: I0313 07:40:04.039814 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:04 crc kubenswrapper[4876]: I0313 07:40:04.040871 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:04 crc kubenswrapper[4876]: I0313 07:40:04.040912 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:04 crc kubenswrapper[4876]: I0313 07:40:04.040929 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:04 crc kubenswrapper[4876]: I0313 07:40:04.040956 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:40:04 crc kubenswrapper[4876]: E0313 07:40:04.044523 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 07:40:04 crc kubenswrapper[4876]: I0313 07:40:04.992928 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:05 crc kubenswrapper[4876]: I0313 07:40:05.992723 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:06 crc kubenswrapper[4876]: I0313 07:40:06.993570 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:07 crc kubenswrapper[4876]: W0313 07:40:07.665266 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:07 crc kubenswrapper[4876]: E0313 07:40:07.665318 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 07:40:07 crc kubenswrapper[4876]: I0313 07:40:07.994782 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:08 crc kubenswrapper[4876]: I0313 07:40:08.994346 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:09 crc kubenswrapper[4876]: I0313 07:40:09.994506 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:10 crc kubenswrapper[4876]: I0313 07:40:10.992977 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:11 crc kubenswrapper[4876]: E0313 07:40:11.034401 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.034922 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.036128 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.036230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.036312 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.036986 4876 scope.go:117] "RemoveContainer" containerID="9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.044858 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.045882 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.045991 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.046070 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.046175 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:40:11 crc kubenswrapper[4876]: E0313 07:40:11.050452 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 07:40:11 crc kubenswrapper[4876]: E0313 07:40:11.147981 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.321099 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.322884 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997"} Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.322996 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.323881 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.323941 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.323962 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:11 crc kubenswrapper[4876]: I0313 07:40:11.994131 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.326953 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.327499 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.329987 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997" exitCode=255 Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.330031 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997"} Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.330082 4876 scope.go:117] "RemoveContainer" containerID="9183604410d5c11d85de9e69c65d37014781da3995145121246ee6d35df50186" Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.330197 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.331356 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.331754 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.331952 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.333640 4876 scope.go:117] "RemoveContainer" containerID="2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997" Mar 13 07:40:12 crc kubenswrapper[4876]: E0313 07:40:12.334135 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.566559 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.586203 4876 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 07:40:12 crc kubenswrapper[4876]: I0313 07:40:12.992405 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.336337 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.442444 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.442630 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.443764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.443799 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.443811 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.444379 4876 scope.go:117] "RemoveContainer" containerID="2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997" Mar 13 07:40:13 crc kubenswrapper[4876]: E0313 07:40:13.444556 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.546500 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:40:13 crc kubenswrapper[4876]: I0313 07:40:13.998320 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:14 crc kubenswrapper[4876]: I0313 07:40:14.340943 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:14 crc kubenswrapper[4876]: I0313 07:40:14.342113 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:14 crc kubenswrapper[4876]: I0313 07:40:14.342151 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:14 crc kubenswrapper[4876]: I0313 07:40:14.342164 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:14 crc kubenswrapper[4876]: I0313 07:40:14.342772 4876 scope.go:117] "RemoveContainer" containerID="2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997" Mar 13 07:40:14 crc kubenswrapper[4876]: E0313 07:40:14.342960 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:40:14 crc kubenswrapper[4876]: W0313 07:40:14.507659 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 13 07:40:14 crc kubenswrapper[4876]: E0313 07:40:14.507706 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 07:40:14 crc kubenswrapper[4876]: I0313 07:40:14.992128 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 07:40:15 crc kubenswrapper[4876]: I0313 07:40:15.870110 4876 csr.go:261] certificate signing request csr-7ctkd is approved, waiting to be issued Mar 13 07:40:15 crc kubenswrapper[4876]: I0313 07:40:15.877593 4876 csr.go:257] certificate signing request csr-7ctkd is issued Mar 13 07:40:15 crc kubenswrapper[4876]: I0313 07:40:15.967367 4876 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 13 07:40:16 crc kubenswrapper[4876]: I0313 07:40:16.833177 4876 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 13 07:40:16 crc kubenswrapper[4876]: I0313 07:40:16.879286 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-25 13:48:50.771974204 +0000 UTC Mar 13 07:40:16 crc kubenswrapper[4876]: I0313 07:40:16.879332 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6894h8m33.892644749s for next certificate rotation Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.050543 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.052313 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.052373 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.052385 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.052544 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.061959 4876 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.062319 4876 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.062350 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.065772 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.065827 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.065844 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.065866 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.065884 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:18Z","lastTransitionTime":"2026-03-13T07:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.085693 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.092285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.092318 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.092329 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.092345 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.092358 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:18Z","lastTransitionTime":"2026-03-13T07:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.101215 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.109869 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.109904 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.109913 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.109926 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.109937 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:18Z","lastTransitionTime":"2026-03-13T07:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.118916 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.126604 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.126635 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.126646 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.126662 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:18 crc kubenswrapper[4876]: I0313 07:40:18.126674 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:18Z","lastTransitionTime":"2026-03-13T07:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.138128 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.138488 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.138538 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.238691 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.339303 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.440307 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.541179 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.642336 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.743307 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.844420 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:18 crc kubenswrapper[4876]: E0313 07:40:18.945305 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.046016 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.146404 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.247429 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.347591 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.448316 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.548581 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.648906 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.749818 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.850344 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:19 crc kubenswrapper[4876]: E0313 07:40:19.951429 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.051792 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: I0313 07:40:20.112032 4876 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.152617 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.253412 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.353834 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.454408 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.555143 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.656289 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.756451 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.857217 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:20 crc kubenswrapper[4876]: E0313 07:40:20.957570 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.058388 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.148338 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.159283 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.259922 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.360048 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.460794 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.561336 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.662390 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.763056 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.863974 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:21 crc kubenswrapper[4876]: E0313 07:40:21.964278 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.065167 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.165480 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.266152 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.367280 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.467641 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.568472 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.668810 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.769324 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.869832 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:22 crc kubenswrapper[4876]: E0313 07:40:22.970976 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.071353 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.171943 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.272648 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.372824 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.473858 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.574967 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.675768 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.776904 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.877374 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:23 crc kubenswrapper[4876]: E0313 07:40:23.977559 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.078372 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.178867 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.279823 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.380622 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.481352 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.581909 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.682382 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.784004 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.884581 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:24 crc kubenswrapper[4876]: E0313 07:40:24.985823 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: I0313 07:40:25.034642 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:25 crc kubenswrapper[4876]: I0313 07:40:25.035874 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:25 crc kubenswrapper[4876]: I0313 07:40:25.035992 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:25 crc kubenswrapper[4876]: I0313 07:40:25.036060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.086879 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.187349 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.288505 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.389607 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.490156 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.590608 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.691884 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.792174 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.892678 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:25 crc kubenswrapper[4876]: E0313 07:40:25.993785 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: I0313 07:40:26.035147 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:26 crc kubenswrapper[4876]: I0313 07:40:26.036484 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:26 crc kubenswrapper[4876]: I0313 07:40:26.036549 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:26 crc kubenswrapper[4876]: I0313 07:40:26.036569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:26 crc kubenswrapper[4876]: I0313 07:40:26.037571 4876 scope.go:117] "RemoveContainer" containerID="2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.037891 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.094011 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.194319 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.294945 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.395723 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.496104 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.597070 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.697560 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.798686 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:26 crc kubenswrapper[4876]: E0313 07:40:26.899093 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.000086 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.100684 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.200840 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.301989 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.402196 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.502721 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.603686 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.704289 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.804802 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:27 crc kubenswrapper[4876]: I0313 07:40:27.824412 4876 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 07:40:27 crc kubenswrapper[4876]: E0313 07:40:27.905876 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.006832 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.107393 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.208306 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.308667 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.409869 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.472769 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.477372 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.477406 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.477422 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.477442 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.477457 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:28Z","lastTransitionTime":"2026-03-13T07:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.494228 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.505711 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.505783 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.505808 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.505836 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.505857 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:28Z","lastTransitionTime":"2026-03-13T07:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.519011 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.523164 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.523516 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.523595 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.523689 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.523776 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:28Z","lastTransitionTime":"2026-03-13T07:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.539138 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.542855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.542927 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.542949 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.542975 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:28 crc kubenswrapper[4876]: I0313 07:40:28.542996 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:28Z","lastTransitionTime":"2026-03-13T07:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.562577 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.562992 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.563089 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.664073 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.764379 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.864826 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:28 crc kubenswrapper[4876]: E0313 07:40:28.965179 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.066094 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.166413 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.266915 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.367382 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.468402 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.569383 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.670367 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.770678 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.870802 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:29 crc kubenswrapper[4876]: E0313 07:40:29.971574 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.072382 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.173321 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.274497 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.374779 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.475726 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.576047 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.677184 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.777516 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.878288 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:30 crc kubenswrapper[4876]: E0313 07:40:30.978669 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.079352 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.149278 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.179966 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.280738 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.381050 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.481793 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.582550 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.683632 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.784570 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.884778 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:31 crc kubenswrapper[4876]: E0313 07:40:31.984976 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.085967 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.186322 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.286509 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.386626 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.487691 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.588705 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.689029 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.789407 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.889707 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:32 crc kubenswrapper[4876]: E0313 07:40:32.990827 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.091350 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.191860 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.292793 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.393538 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.493837 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.594644 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.695821 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.796366 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.896810 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:33 crc kubenswrapper[4876]: E0313 07:40:33.997918 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.098071 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.198981 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.299800 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.400136 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.500518 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.601683 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.701987 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.803050 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:34 crc kubenswrapper[4876]: E0313 07:40:34.904112 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.004826 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.106012 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.206450 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.307572 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.408290 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.509403 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.610268 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.710631 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.810765 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:35 crc kubenswrapper[4876]: E0313 07:40:35.911337 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.011679 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.111854 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.212345 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.312700 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.413462 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.514485 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.614879 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.715729 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.816634 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:36 crc kubenswrapper[4876]: E0313 07:40:36.917427 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.017999 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.118150 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.218979 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.320129 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.420441 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.521253 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.622274 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.723184 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.824148 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:37 crc kubenswrapper[4876]: E0313 07:40:37.924286 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.025008 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.125290 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.226151 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.327308 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.428491 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.529086 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.629678 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.730055 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.831175 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.931853 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.956053 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.960868 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.960924 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.960934 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.960949 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.960961 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:38Z","lastTransitionTime":"2026-03-13T07:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.975138 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.980855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.980941 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.980956 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.980983 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:38 crc kubenswrapper[4876]: I0313 07:40:38.980997 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:38Z","lastTransitionTime":"2026-03-13T07:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:38 crc kubenswrapper[4876]: E0313 07:40:38.997302 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.002202 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.002296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.002316 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.002345 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.002368 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:39Z","lastTransitionTime":"2026-03-13T07:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.020011 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.024549 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.024625 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.024663 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.024700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.024726 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:39Z","lastTransitionTime":"2026-03-13T07:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.034999 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.035919 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.036186 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.036233 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.038041 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.038139 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.038163 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:39 crc kubenswrapper[4876]: I0313 07:40:39.039665 4876 scope.go:117] "RemoveContainer" containerID="2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.040157 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.136937 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.237568 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.338637 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.439818 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.540064 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.640524 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.741297 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.841663 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:39 crc kubenswrapper[4876]: E0313 07:40:39.942475 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.042818 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.144112 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.245130 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.345512 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.446020 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.546832 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.647160 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.748321 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.849430 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:40 crc kubenswrapper[4876]: E0313 07:40:40.949605 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.049923 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.150277 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.150348 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.250447 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.350565 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.451380 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.551812 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.652863 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.753774 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.854639 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:41 crc kubenswrapper[4876]: E0313 07:40:41.955109 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: I0313 07:40:42.034488 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 07:40:42 crc kubenswrapper[4876]: I0313 07:40:42.035739 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:42 crc kubenswrapper[4876]: I0313 07:40:42.035832 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:42 crc kubenswrapper[4876]: I0313 07:40:42.035858 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.056041 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.156831 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.257949 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.359126 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.459999 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.560686 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.661082 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.761587 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.862422 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:42 crc kubenswrapper[4876]: E0313 07:40:42.962944 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.063429 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.164154 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.264474 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.364757 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.465257 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.565551 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.666425 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.767308 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.867695 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:43 crc kubenswrapper[4876]: E0313 07:40:43.968445 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.069512 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.170125 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.270487 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.370862 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.471305 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.571722 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.671906 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.772057 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.873029 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:44 crc kubenswrapper[4876]: E0313 07:40:44.973689 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.074370 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.175007 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.275184 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.375736 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.476707 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.577494 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.678014 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.779154 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.879297 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:45 crc kubenswrapper[4876]: E0313 07:40:45.979469 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.079963 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.180977 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.282362 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.382516 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.483005 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.584099 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.684305 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.785457 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.885650 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:46 crc kubenswrapper[4876]: E0313 07:40:46.986055 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.087011 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.187282 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.287850 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.388110 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.488874 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.589721 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.690729 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.790970 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.891425 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:47 crc kubenswrapper[4876]: E0313 07:40:47.991550 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.091903 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.192598 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.292917 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.393840 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.494343 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.594866 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.695055 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.796032 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.897072 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:48 crc kubenswrapper[4876]: E0313 07:40:48.997281 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.098291 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.198656 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.262587 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.266562 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.266609 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.266634 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.266664 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.266686 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:49Z","lastTransitionTime":"2026-03-13T07:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.278008 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.281713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.281770 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.281784 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.281800 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.281814 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:49Z","lastTransitionTime":"2026-03-13T07:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.290395 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.294630 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.294705 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.294731 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.294757 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.294775 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:49Z","lastTransitionTime":"2026-03-13T07:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.307574 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.312445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.312509 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.312524 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.312549 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:49 crc kubenswrapper[4876]: I0313 07:40:49.312563 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:49Z","lastTransitionTime":"2026-03-13T07:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.322667 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.322784 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.322818 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.423683 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.524718 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.625650 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.726094 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.827080 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:49 crc kubenswrapper[4876]: E0313 07:40:49.928105 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.028918 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.129452 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.230390 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.330873 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.431627 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.531767 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.632503 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.733488 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.834332 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:50 crc kubenswrapper[4876]: E0313 07:40:50.934628 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.034936 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.135906 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.151143 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.236453 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.337567 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.437974 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.538425 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.639171 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.739952 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.840965 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:51 crc kubenswrapper[4876]: E0313 07:40:51.941052 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.041310 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.142410 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.242686 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.343293 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.443833 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.545008 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.645917 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.746296 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.846486 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:52 crc kubenswrapper[4876]: E0313 07:40:52.947014 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:53 crc kubenswrapper[4876]: E0313 07:40:53.047282 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:53 crc kubenswrapper[4876]: E0313 07:40:53.148348 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:53 crc kubenswrapper[4876]: E0313 07:40:53.248687 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:53 crc kubenswrapper[4876]: E0313 07:40:53.349407 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:53 crc kubenswrapper[4876]: E0313 07:40:53.449815 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:53 crc kubenswrapper[4876]: E0313 07:40:53.550511 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:53 crc kubenswrapper[4876]: E0313 07:40:53.651453 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.738676 4876 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.755095 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.755530 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.755781 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.756084 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.756425 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:53Z","lastTransitionTime":"2026-03-13T07:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.858755 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.858788 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.858800 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.858813 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.858823 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:53Z","lastTransitionTime":"2026-03-13T07:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.962255 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.962304 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.962315 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.962331 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:53 crc kubenswrapper[4876]: I0313 07:40:53.962342 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:53Z","lastTransitionTime":"2026-03-13T07:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.050368 4876 scope.go:117] "RemoveContainer" containerID="2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.066148 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.066217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.066275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.066307 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.066345 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.168760 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.168805 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.168815 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.168831 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.168841 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.173692 4876 apiserver.go:52] "Watching apiserver" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.177220 4876 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.177767 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-krrjs","openshift-multus/multus-st667","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k","openshift-ovn-kubernetes/ovnkube-node-g4xkn","openshift-image-registry/node-ca-sz689","openshift-multus/multus-additional-cni-plugins-gqsld","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-machine-config-operator/machine-config-daemon-r9cl2","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-multus/network-metrics-daemon-zrbh4"] Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.178415 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.178447 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.178524 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.178651 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.178697 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.178791 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.178847 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.178995 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.179062 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-krrjs" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.179092 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.179116 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.179461 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.179506 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.179533 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.179576 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.180060 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.180100 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.180816 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.181456 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.181513 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.181619 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.182778 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.182818 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.182973 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.183124 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.183279 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.183336 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.183710 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.184282 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.184526 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.184743 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.184774 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.184943 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.184959 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.185055 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.185304 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.185367 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.185444 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.185573 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.185728 4876 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.186163 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.186203 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.186707 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.186803 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.186856 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.186898 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.186818 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.187351 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.187380 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.188412 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.189015 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.189179 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.190488 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.187511 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.190850 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.192790 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.203903 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.205931 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206047 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206133 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206215 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206310 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206380 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206516 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206598 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206350 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206450 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206475 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.206639 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:40:54.706620266 +0000 UTC m=+114.377399348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207018 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207092 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206731 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.206886 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207157 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207100 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207109 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207397 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207474 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207504 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207526 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207549 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207570 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207592 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207615 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207635 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207657 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207678 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207701 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207724 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207747 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207768 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207822 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207310 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207835 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207816 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207845 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207879 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207903 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207924 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207946 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.207968 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208030 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208055 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208079 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208106 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208130 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208153 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208174 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208195 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208218 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208300 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208321 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208348 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208369 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208389 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208412 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208433 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208454 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208474 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208495 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208517 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208541 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208565 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208589 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208613 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208637 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208662 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208683 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208707 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208730 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208751 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208769 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208792 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208813 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208839 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208861 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208885 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208905 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208926 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208947 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208969 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208991 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209013 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209036 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209059 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209082 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209109 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209134 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209158 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209180 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209202 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209223 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209275 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209299 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209323 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209343 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209363 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209385 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209406 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209427 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209452 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209474 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209498 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209524 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209547 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209570 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209594 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209615 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209637 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209657 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209678 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209698 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209722 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209746 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209768 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209790 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209817 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209838 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209859 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209887 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209907 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209931 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209956 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209980 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210003 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210025 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210051 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210074 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210097 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210121 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210146 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210168 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210194 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210217 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210257 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210298 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210323 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210347 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210370 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210395 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210418 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210440 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210466 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210489 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210512 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210537 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210560 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210583 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210607 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210629 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210653 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210676 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210702 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210724 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210753 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210776 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210801 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210827 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210850 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210874 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210896 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210919 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210998 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211019 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211045 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211070 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211253 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211282 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211307 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211331 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211354 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211379 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211403 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211429 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211452 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211475 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211497 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211519 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211546 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211571 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211600 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211623 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211647 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211669 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211690 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211714 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211741 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211766 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211814 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211836 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211861 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211886 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211911 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211937 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211960 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211982 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212044 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212079 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212105 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212130 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212158 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212187 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212211 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212363 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-kubelet\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212399 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-slash\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212420 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-log-socket\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212443 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s45rv\" (UniqueName: \"kubernetes.io/projected/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-kube-api-access-s45rv\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212474 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212651 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212681 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cni-binary-copy\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212734 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212755 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-env-overrides\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212777 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovn-node-metrics-cert\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212800 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-etc-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212825 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cdd10539-ff2d-4055-86eb-d422e532bf66-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212849 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-daemon-config\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212871 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-multus-certs\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212899 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212928 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212953 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cdd10539-ff2d-4055-86eb-d422e532bf66-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212976 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/182bf6da-086a-4b4e-b69c-411b5f6e8593-serviceca\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212996 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b230d31c-acdd-44a2-85b5-30e786f515c9-hosts-file\") pod \"node-resolver-krrjs\" (UID: \"b230d31c-acdd-44a2-85b5-30e786f515c9\") " pod="openshift-dns/node-resolver-krrjs" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213016 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-cni-multus\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213036 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-hostroot\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213059 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-netns\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213082 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgt4l\" (UniqueName: \"kubernetes.io/projected/cdd10539-ff2d-4055-86eb-d422e532bf66-kube-api-access-jgt4l\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213109 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213132 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/182bf6da-086a-4b4e-b69c-411b5f6e8593-host\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213154 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-socket-dir-parent\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213178 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-netns\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213200 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-cni-bin\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213221 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213266 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-systemd-units\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213288 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-config\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213314 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0a6f71e5-2091-4386-b559-bba70bc45972-rootfs\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208192 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208413 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208526 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208552 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208619 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208811 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208814 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.208830 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209071 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209115 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209394 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209457 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209499 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209503 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.209908 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210075 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210084 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210418 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210450 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210634 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210736 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.210912 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211148 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211294 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211318 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211391 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.211696 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212065 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212158 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213656 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212198 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212221 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212355 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212497 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212562 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212666 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212682 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212813 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212874 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.212981 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213145 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213169 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213013 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213378 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213541 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213872 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.214008 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.214204 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.214296 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.214613 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.214688 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.214816 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.214861 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215008 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215041 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215375 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215419 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215489 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215549 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215669 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215801 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215965 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215972 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.215989 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.216297 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.216533 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.216696 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.216760 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.216844 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.216859 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.216135 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217188 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217332 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.216567 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217568 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217559 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217626 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217692 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217821 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217918 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217948 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.217957 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.218039 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.218262 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.218697 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.218839 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.218905 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219027 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219046 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219090 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219219 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219294 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219360 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219394 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219584 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219610 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219925 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.219959 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.220065 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.220104 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.220374 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.220397 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.220471 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.220916 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221145 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221193 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221333 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221437 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.213444 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-cni-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221743 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjs89\" (UniqueName: \"kubernetes.io/projected/f36294fb-98b9-48be-8237-beb6484e0cb6-kube-api-access-bjs89\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221781 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221808 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-cnibin\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221830 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-os-release\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221853 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-netd\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221884 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221908 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a6f71e5-2091-4386-b559-bba70bc45972-mcd-auth-proxy-config\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221934 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-system-cni-dir\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221957 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-ovn-kubernetes\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221977 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cdd10539-ff2d-4055-86eb-d422e532bf66-env-overrides\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222002 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222028 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222051 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-systemd\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222073 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-var-lib-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222093 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-ovn\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222113 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-node-log\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222132 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-bin\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222154 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222175 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222194 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twfww\" (UniqueName: \"kubernetes.io/projected/b230d31c-acdd-44a2-85b5-30e786f515c9-kube-api-access-twfww\") pod \"node-resolver-krrjs\" (UID: \"b230d31c-acdd-44a2-85b5-30e786f515c9\") " pod="openshift-dns/node-resolver-krrjs" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222216 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxrm7\" (UniqueName: \"kubernetes.io/projected/182bf6da-086a-4b4e-b69c-411b5f6e8593-kube-api-access-sxrm7\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222252 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-system-cni-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222272 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8adb2a69-d0ef-4efc-813d-77fdf9713d50-cni-binary-copy\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222296 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-os-release\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222323 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221683 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.221915 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.218427 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222229 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222370 4876 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222434 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222541 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222551 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxm4r\" (UniqueName: \"kubernetes.io/projected/8adb2a69-d0ef-4efc-813d-77fdf9713d50-kube-api-access-sxm4r\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222621 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.222864 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.223095 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.223146 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.223505 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.223790 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.224211 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.224216 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.224632 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.224649 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.224746 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.224802 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.225091 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.225429 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226822 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226893 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226923 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226952 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-script-lib\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226977 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-k8s-cni-cncf-io\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226998 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.227025 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smtrk\" (UniqueName: \"kubernetes.io/projected/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-kube-api-access-smtrk\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.227059 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.227160 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.227820 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-kubelet\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.228002 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0a6f71e5-2091-4386-b559-bba70bc45972-proxy-tls\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.228293 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmnf2\" (UniqueName: \"kubernetes.io/projected/0a6f71e5-2091-4386-b559-bba70bc45972-kube-api-access-xmnf2\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.228358 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-etc-kubernetes\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.228383 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cnibin\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.228406 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-conf-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.225513 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.225618 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.225756 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.226165 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.228892 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:54.728869756 +0000 UTC m=+114.399648808 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.229289 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.229360 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:54.729343479 +0000 UTC m=+114.400122461 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226267 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226343 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226383 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226420 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226554 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226579 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226748 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226858 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226881 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.226938 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.227073 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.227259 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.227303 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.228607 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.228668 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.228703 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.229811 4876 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.229829 4876 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230059 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230090 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230104 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230116 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230128 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230140 4876 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230153 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230165 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230176 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230187 4876 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230198 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230211 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230223 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230278 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230291 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230303 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230316 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230327 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230339 4876 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230352 4876 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230363 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230374 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230386 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230397 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230408 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230420 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230431 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230442 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230455 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230468 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230479 4876 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230490 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230501 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230512 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230525 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230537 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230549 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230560 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230571 4876 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230582 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230593 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230605 4876 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230616 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230628 4876 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230640 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230651 4876 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230661 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230711 4876 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230725 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230736 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230749 4876 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230760 4876 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230797 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230811 4876 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230821 4876 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230832 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230843 4876 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230878 4876 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230889 4876 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230900 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230911 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230922 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230957 4876 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230970 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230981 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.230993 4876 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231005 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231039 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231052 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231064 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231075 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231089 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231128 4876 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231141 4876 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231151 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231163 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231176 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231208 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231220 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231230 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231268 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231280 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231291 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231302 4876 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231312 4876 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231345 4876 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231357 4876 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231369 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231380 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231390 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231402 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231413 4876 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231425 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231435 4876 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231445 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231456 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231467 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231479 4876 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231490 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231521 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231534 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231548 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231559 4876 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231571 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231582 4876 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232771 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232791 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232807 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232821 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232834 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232845 4876 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232857 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232868 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232880 4876 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232892 4876 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232903 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.231635 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.232909 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.240652 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.240686 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.240702 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.240764 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:54.740745932 +0000 UTC m=+114.411524914 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.242148 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.242174 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.242186 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.242224 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:54.742213362 +0000 UTC m=+114.412992344 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.245765 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246177 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246252 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246298 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246326 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246407 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246453 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246477 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246541 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246569 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246675 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246686 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.246991 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.248014 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.248401 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.248435 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.248754 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.248821 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.249640 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.249789 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.249824 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.249979 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.250405 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.251167 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.251925 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.253469 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.253575 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.253645 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.253748 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.253986 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.254035 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.254164 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.254749 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.255688 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.256443 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.256483 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.256633 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.256730 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.258928 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.259703 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.260402 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.268500 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.271337 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.272758 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.272804 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.272814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.272826 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.272836 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.275478 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.280327 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.282136 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.286993 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.289662 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.296603 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.298975 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.307151 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.317574 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.327967 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333346 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-ovn-kubernetes\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333391 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cdd10539-ff2d-4055-86eb-d422e532bf66-env-overrides\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333426 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-bin\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333467 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-systemd\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333494 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-var-lib-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333519 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-ovn\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333540 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-node-log\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333563 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333588 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twfww\" (UniqueName: \"kubernetes.io/projected/b230d31c-acdd-44a2-85b5-30e786f515c9-kube-api-access-twfww\") pod \"node-resolver-krrjs\" (UID: \"b230d31c-acdd-44a2-85b5-30e786f515c9\") " pod="openshift-dns/node-resolver-krrjs" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333612 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxrm7\" (UniqueName: \"kubernetes.io/projected/182bf6da-086a-4b4e-b69c-411b5f6e8593-kube-api-access-sxrm7\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333633 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-system-cni-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333658 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8adb2a69-d0ef-4efc-813d-77fdf9713d50-cni-binary-copy\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333683 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-os-release\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333782 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-node-log\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333741 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxm4r\" (UniqueName: \"kubernetes.io/projected/8adb2a69-d0ef-4efc-813d-77fdf9713d50-kube-api-access-sxm4r\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333826 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-ovn-kubernetes\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333844 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333859 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-systemd\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333954 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-ovn\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333783 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-bin\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.333969 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.334024 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-var-lib-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.334016 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.334080 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-script-lib\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.334444 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cdd10539-ff2d-4055-86eb-d422e532bf66-env-overrides\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.334517 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-system-cni-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.334547 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.334731 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8adb2a69-d0ef-4efc-813d-77fdf9713d50-cni-binary-copy\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.335450 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-os-release\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.335450 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-k8s-cni-cncf-io\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.335526 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.335539 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.335590 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smtrk\" (UniqueName: \"kubernetes.io/projected/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-kube-api-access-smtrk\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.335631 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-k8s-cni-cncf-io\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.335875 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-kubelet\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336016 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0a6f71e5-2091-4386-b559-bba70bc45972-proxy-tls\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336049 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmnf2\" (UniqueName: \"kubernetes.io/projected/0a6f71e5-2091-4386-b559-bba70bc45972-kube-api-access-xmnf2\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336458 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-script-lib\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336515 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-kubelet\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336543 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-etc-kubernetes\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336584 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-etc-kubernetes\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336594 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cnibin\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-conf-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336777 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-kubelet\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336810 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cni-binary-copy\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336838 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-slash\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336861 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-log-socket\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336883 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s45rv\" (UniqueName: \"kubernetes.io/projected/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-kube-api-access-s45rv\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336932 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336953 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-env-overrides\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336964 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-kubelet\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336979 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovn-node-metrics-cert\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.336630 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cnibin\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337030 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-etc-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337059 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cdd10539-ff2d-4055-86eb-d422e532bf66-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337087 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-daemon-config\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337115 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-multus-certs\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337161 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cdd10539-ff2d-4055-86eb-d422e532bf66-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337189 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/182bf6da-086a-4b4e-b69c-411b5f6e8593-serviceca\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337217 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b230d31c-acdd-44a2-85b5-30e786f515c9-hosts-file\") pod \"node-resolver-krrjs\" (UID: \"b230d31c-acdd-44a2-85b5-30e786f515c9\") " pod="openshift-dns/node-resolver-krrjs" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337260 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-cni-multus\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337282 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-hostroot\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337308 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-netns\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337333 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-cni-bin\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337361 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-netns\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337388 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgt4l\" (UniqueName: \"kubernetes.io/projected/cdd10539-ff2d-4055-86eb-d422e532bf66-kube-api-access-jgt4l\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337454 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/182bf6da-086a-4b4e-b69c-411b5f6e8593-host\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337478 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-socket-dir-parent\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337503 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337533 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-systemd-units\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337539 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337571 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-config\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337601 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0a6f71e5-2091-4386-b559-bba70bc45972-rootfs\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337613 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-cni-multus\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337628 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-cni-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337656 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjs89\" (UniqueName: \"kubernetes.io/projected/f36294fb-98b9-48be-8237-beb6484e0cb6-kube-api-access-bjs89\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337668 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-conf-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337704 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337717 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-etc-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337732 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-cnibin\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337759 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-os-release\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337791 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-netd\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337823 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a6f71e5-2091-4386-b559-bba70bc45972-mcd-auth-proxy-config\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337848 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-system-cni-dir\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337967 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337989 4876 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338007 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338026 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338038 4876 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338050 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338063 4876 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338081 4876 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338093 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338106 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338118 4876 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338133 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338144 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338156 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338154 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-cni-binary-copy\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338170 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338182 4876 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338195 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338209 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-slash\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338213 4876 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338260 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-system-cni-dir\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338273 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338297 4876 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338313 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338326 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338337 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338357 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338376 4876 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338388 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338400 4876 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338414 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338423 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338438 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338447 4876 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338459 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338472 4876 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338477 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0a6f71e5-2091-4386-b559-bba70bc45972-rootfs\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338483 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338501 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cdd10539-ff2d-4055-86eb-d422e532bf66-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338513 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338546 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-cnibin\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338562 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338536 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-cni-dir\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.337629 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-log-socket\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338588 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-os-release\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338589 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338611 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338625 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338822 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-config\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338855 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338876 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-netns\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338882 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-hostroot\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338901 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/182bf6da-086a-4b4e-b69c-411b5f6e8593-host\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338932 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-var-lib-cni-bin\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338947 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-systemd-units\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.338934 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.338965 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-netns\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.339003 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs podName:f36294fb-98b9-48be-8237-beb6484e0cb6 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:54.838983824 +0000 UTC m=+114.509762806 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs") pod "network-metrics-daemon-zrbh4" (UID: "f36294fb-98b9-48be-8237-beb6484e0cb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339012 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-socket-dir-parent\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339055 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-netd\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339114 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8adb2a69-d0ef-4efc-813d-77fdf9713d50-host-run-multus-certs\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339515 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-openvswitch\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339626 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b230d31c-acdd-44a2-85b5-30e786f515c9-hosts-file\") pod \"node-resolver-krrjs\" (UID: \"b230d31c-acdd-44a2-85b5-30e786f515c9\") " pod="openshift-dns/node-resolver-krrjs" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339670 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-env-overrides\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339932 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339962 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.339981 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340086 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340111 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340133 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340149 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340162 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340175 4876 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340192 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340204 4876 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340216 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340230 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340261 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340273 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340286 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340298 4876 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340308 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340318 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340330 4876 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340358 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340371 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340384 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340401 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340415 4876 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340428 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340441 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340457 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340469 4876 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340480 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340492 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340509 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340522 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340541 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340557 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340575 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340588 4876 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.340638 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a6f71e5-2091-4386-b559-bba70bc45972-mcd-auth-proxy-config\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.348325 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.350470 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8adb2a69-d0ef-4efc-813d-77fdf9713d50-multus-daemon-config\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.351896 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twfww\" (UniqueName: \"kubernetes.io/projected/b230d31c-acdd-44a2-85b5-30e786f515c9-kube-api-access-twfww\") pod \"node-resolver-krrjs\" (UID: \"b230d31c-acdd-44a2-85b5-30e786f515c9\") " pod="openshift-dns/node-resolver-krrjs" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.352588 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/182bf6da-086a-4b4e-b69c-411b5f6e8593-serviceca\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.353139 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0a6f71e5-2091-4386-b559-bba70bc45972-proxy-tls\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.353997 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmnf2\" (UniqueName: \"kubernetes.io/projected/0a6f71e5-2091-4386-b559-bba70bc45972-kube-api-access-xmnf2\") pod \"machine-config-daemon-r9cl2\" (UID: \"0a6f71e5-2091-4386-b559-bba70bc45972\") " pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.355406 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxm4r\" (UniqueName: \"kubernetes.io/projected/8adb2a69-d0ef-4efc-813d-77fdf9713d50-kube-api-access-sxm4r\") pod \"multus-st667\" (UID: \"8adb2a69-d0ef-4efc-813d-77fdf9713d50\") " pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.355792 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smtrk\" (UniqueName: \"kubernetes.io/projected/153920eb-fdb1-4e94-85ca-f09b5f01c2aa-kube-api-access-smtrk\") pod \"multus-additional-cni-plugins-gqsld\" (UID: \"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\") " pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.356049 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cdd10539-ff2d-4055-86eb-d422e532bf66-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.357640 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjs89\" (UniqueName: \"kubernetes.io/projected/f36294fb-98b9-48be-8237-beb6484e0cb6-kube-api-access-bjs89\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.359157 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgt4l\" (UniqueName: \"kubernetes.io/projected/cdd10539-ff2d-4055-86eb-d422e532bf66-kube-api-access-jgt4l\") pod \"ovnkube-control-plane-749d76644c-59l8k\" (UID: \"cdd10539-ff2d-4055-86eb-d422e532bf66\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.359571 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovn-node-metrics-cert\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.360381 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.363968 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s45rv\" (UniqueName: \"kubernetes.io/projected/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-kube-api-access-s45rv\") pod \"ovnkube-node-g4xkn\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.365046 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxrm7\" (UniqueName: \"kubernetes.io/projected/182bf6da-086a-4b4e-b69c-411b5f6e8593-kube-api-access-sxrm7\") pod \"node-ca-sz689\" (UID: \"182bf6da-086a-4b4e-b69c-411b5f6e8593\") " pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.368467 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.375636 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.375677 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.375689 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.375724 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.375738 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.380834 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.393035 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.445715 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.447406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.447699 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.448672 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fa1ffd9770cea5029987347a06d913c70e6ab58894d42f9f9a5169981983b52a"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.457697 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.468145 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.478349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.478374 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.478381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.478392 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.478401 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.481340 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.489603 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.502930 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.506098 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.512179 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.524330 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.535120 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.544463 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.555086 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.560388 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gqsld" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.563074 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.569233 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-st667" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.570656 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.576162 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.580094 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.580127 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.580139 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.580158 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.580256 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.580388 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: W0313 07:40:54.587676 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8adb2a69_d0ef_4efc_813d_77fdf9713d50.slice/crio-8dad2cdcef349a8b4b33f27b70807c1c19608144d6e4f56036c82861e17c35ae WatchSource:0}: Error finding container 8dad2cdcef349a8b4b33f27b70807c1c19608144d6e4f56036c82861e17c35ae: Status 404 returned error can't find the container with id 8dad2cdcef349a8b4b33f27b70807c1c19608144d6e4f56036c82861e17c35ae Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.590921 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-sz689" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.591033 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.598426 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 07:40:54 crc kubenswrapper[4876]: W0313 07:40:54.606338 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a2d8ff1_99c6_44ed_bb4d_cd753c9126fd.slice/crio-0e9da90db3918b231f3e15e05b6b39eab6b349dc5e7ff7cbb9584c79f673692e WatchSource:0}: Error finding container 0e9da90db3918b231f3e15e05b6b39eab6b349dc5e7ff7cbb9584c79f673692e: Status 404 returned error can't find the container with id 0e9da90db3918b231f3e15e05b6b39eab6b349dc5e7ff7cbb9584c79f673692e Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.606639 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.618208 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" Mar 13 07:40:54 crc kubenswrapper[4876]: W0313 07:40:54.618647 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod182bf6da_086a_4b4e_b69c_411b5f6e8593.slice/crio-d19bf75c94f33e5fee149489df4a80bee16774e694e372b8f10b108c0452aa63 WatchSource:0}: Error finding container d19bf75c94f33e5fee149489df4a80bee16774e694e372b8f10b108c0452aa63: Status 404 returned error can't find the container with id d19bf75c94f33e5fee149489df4a80bee16774e694e372b8f10b108c0452aa63 Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.628884 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 07:40:54 crc kubenswrapper[4876]: W0313 07:40:54.637172 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdd10539_ff2d_4055_86eb_d422e532bf66.slice/crio-d67f779b9de9446512ef36aad505a74dae836fd9fa86c1f5f68c794bafc0a187 WatchSource:0}: Error finding container d67f779b9de9446512ef36aad505a74dae836fd9fa86c1f5f68c794bafc0a187: Status 404 returned error can't find the container with id d67f779b9de9446512ef36aad505a74dae836fd9fa86c1f5f68c794bafc0a187 Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.647624 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-krrjs" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.682663 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.682697 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.682706 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.682719 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.682728 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: W0313 07:40:54.693987 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb230d31c_acdd_44a2_85b5_30e786f515c9.slice/crio-7f6c7a7b015cc0dade1a0130da4dc8b9f9d58a5e1a676e42d204c2c7394ffafd WatchSource:0}: Error finding container 7f6c7a7b015cc0dade1a0130da4dc8b9f9d58a5e1a676e42d204c2c7394ffafd: Status 404 returned error can't find the container with id 7f6c7a7b015cc0dade1a0130da4dc8b9f9d58a5e1a676e42d204c2c7394ffafd Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.745205 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.745324 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:40:55.745300378 +0000 UTC m=+115.416079370 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.745472 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.745946 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.746115 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:55.745992286 +0000 UTC m=+115.416771268 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.746371 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.746393 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.746574 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.749297 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:55.749274917 +0000 UTC m=+115.420053909 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.749348 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.749436 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.749499 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.749619 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.749686 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:55.749669617 +0000 UTC m=+115.420448599 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.749741 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.749758 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.749769 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.749831 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:55.749818471 +0000 UTC m=+115.420597513 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.789193 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.789227 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.789278 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.789294 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.789304 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.850167 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.850318 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: E0313 07:40:54.850369 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs podName:f36294fb-98b9-48be-8237-beb6484e0cb6 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:55.850352926 +0000 UTC m=+115.521131908 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs") pod "network-metrics-daemon-zrbh4" (UID: "f36294fb-98b9-48be-8237-beb6484e0cb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.893565 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.893917 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.893930 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.893946 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.893960 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.995621 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.995658 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.995667 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.995679 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:54 crc kubenswrapper[4876]: I0313 07:40:54.995688 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:54Z","lastTransitionTime":"2026-03-13T07:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.039154 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.039878 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.040565 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.041177 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.041797 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.043186 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.043785 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.044465 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.045463 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.045945 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.046822 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.047595 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.048435 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.048927 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.049812 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.050339 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.050892 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.051809 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.052381 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.053379 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.054670 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.055821 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.056768 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.057461 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.058268 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.058873 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.060027 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.060642 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.061703 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.062163 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.062671 4876 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.062769 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.065284 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.065860 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.066801 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.068500 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.069129 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.070028 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.070735 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.071783 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.072439 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.073056 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.074100 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.075803 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.076412 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.077635 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.079445 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.081486 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.082061 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.083175 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.083774 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.084470 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.085615 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.086173 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.098085 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.098138 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.098149 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.098171 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.098183 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.200963 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.201001 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.201011 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.201024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.201033 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.303347 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.303388 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.303397 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.303417 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.303427 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.405895 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.405951 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.405968 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.405984 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.405995 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.453801 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.453845 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.456156 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-krrjs" event={"ID":"b230d31c-acdd-44a2-85b5-30e786f515c9","Type":"ContainerStarted","Data":"ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.456189 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-krrjs" event={"ID":"b230d31c-acdd-44a2-85b5-30e786f515c9","Type":"ContainerStarted","Data":"7f6c7a7b015cc0dade1a0130da4dc8b9f9d58a5e1a676e42d204c2c7394ffafd"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.457669 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-sz689" event={"ID":"182bf6da-086a-4b4e-b69c-411b5f6e8593","Type":"ContainerStarted","Data":"048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.457695 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-sz689" event={"ID":"182bf6da-086a-4b4e-b69c-411b5f6e8593","Type":"ContainerStarted","Data":"d19bf75c94f33e5fee149489df4a80bee16774e694e372b8f10b108c0452aa63"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.459431 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf" exitCode=0 Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.459449 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.459472 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"0e9da90db3918b231f3e15e05b6b39eab6b349dc5e7ff7cbb9584c79f673692e"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.462468 4876 generic.go:334] "Generic (PLEG): container finished" podID="153920eb-fdb1-4e94-85ca-f09b5f01c2aa" containerID="b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251" exitCode=0 Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.462538 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" event={"ID":"153920eb-fdb1-4e94-85ca-f09b5f01c2aa","Type":"ContainerDied","Data":"b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.462593 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" event={"ID":"153920eb-fdb1-4e94-85ca-f09b5f01c2aa","Type":"ContainerStarted","Data":"f8b2f45dd1299848035e42bf345f1acb9c45cfd4ee71df6ab94aa2090270d0e7"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.465902 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.465954 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.465980 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"19ee9b984d5463045956bc584bbda8bff782b0beddac0edc11bca432f45bad9a"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.468102 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.468174 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"cc9a47e6962db92ab4f454656a0c550755ae53eca499416432ff6ffdeedca663"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.470953 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" event={"ID":"cdd10539-ff2d-4055-86eb-d422e532bf66","Type":"ContainerStarted","Data":"35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.470997 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" event={"ID":"cdd10539-ff2d-4055-86eb-d422e532bf66","Type":"ContainerStarted","Data":"20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.471008 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" event={"ID":"cdd10539-ff2d-4055-86eb-d422e532bf66","Type":"ContainerStarted","Data":"d67f779b9de9446512ef36aad505a74dae836fd9fa86c1f5f68c794bafc0a187"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.472286 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5a5b48cbd9659f95fe942c167159d2a0000a25630f70a0b9102bacdcdf9c8224"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.473712 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.473974 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-st667" event={"ID":"8adb2a69-d0ef-4efc-813d-77fdf9713d50","Type":"ContainerStarted","Data":"77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.474024 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-st667" event={"ID":"8adb2a69-d0ef-4efc-813d-77fdf9713d50","Type":"ContainerStarted","Data":"8dad2cdcef349a8b4b33f27b70807c1c19608144d6e4f56036c82861e17c35ae"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.493706 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.505886 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.508903 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.508945 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.508959 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.508978 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.508991 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.522273 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.536361 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.547985 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.566455 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.584199 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.604642 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.613713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.613736 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.613744 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.613756 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.613765 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.626937 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.640756 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.653650 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.669256 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.683577 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.702160 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.715384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.715438 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.715447 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.715462 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.715471 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.716668 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.732627 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.744885 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.758171 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.759400 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.759536 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759560 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:40:57.75954009 +0000 UTC m=+117.430319072 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.759584 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.759638 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.759677 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759788 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759811 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759839 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759845 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759921 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759939 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759858 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:57.759850649 +0000 UTC m=+117.430629631 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.759861 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.760060 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.760092 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:57.760025414 +0000 UTC m=+117.430804396 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.760129 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:57.760116617 +0000 UTC m=+117.430895709 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.760156 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:57.760145737 +0000 UTC m=+117.430924829 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.776763 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.795623 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.813701 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.819360 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.819396 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.819407 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.819423 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.819434 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.829414 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.842341 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.856032 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.860375 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.860671 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: E0313 07:40:55.860779 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs podName:f36294fb-98b9-48be-8237-beb6484e0cb6 nodeName:}" failed. No retries permitted until 2026-03-13 07:40:57.860749974 +0000 UTC m=+117.531528956 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs") pod "network-metrics-daemon-zrbh4" (UID: "f36294fb-98b9-48be-8237-beb6484e0cb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.871098 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.888026 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.905942 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.921017 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.922020 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.922070 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.922083 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.922106 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.922118 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:55Z","lastTransitionTime":"2026-03-13T07:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:55 crc kubenswrapper[4876]: I0313 07:40:55.931943 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:55Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.025007 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.025048 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.025056 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.025072 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.025084 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.035291 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.035313 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.035297 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:56 crc kubenswrapper[4876]: E0313 07:40:56.035399 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.035291 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:56 crc kubenswrapper[4876]: E0313 07:40:56.035503 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:40:56 crc kubenswrapper[4876]: E0313 07:40:56.035605 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:40:56 crc kubenswrapper[4876]: E0313 07:40:56.035671 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.127840 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.128189 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.128201 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.128215 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.128224 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.230426 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.230460 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.230470 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.230484 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.230496 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.333165 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.333201 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.333212 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.333229 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.333261 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.435849 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.435902 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.435913 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.435928 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.435956 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.480187 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.481362 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.481482 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.481571 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.482020 4876 generic.go:334] "Generic (PLEG): container finished" podID="153920eb-fdb1-4e94-85ca-f09b5f01c2aa" containerID="a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8" exitCode=0 Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.482091 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" event={"ID":"153920eb-fdb1-4e94-85ca-f09b5f01c2aa","Type":"ContainerDied","Data":"a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.505327 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.523160 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.537886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.537927 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.537938 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.537952 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.537961 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.539598 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.556554 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.579689 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.591030 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.602957 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.620437 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.638132 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.642046 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.642109 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.642125 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.642150 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.642171 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.661208 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.674472 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.695854 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.715832 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.733028 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.745051 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.745286 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.745374 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.745443 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.745503 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.745896 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:56Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.848266 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.848303 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.848314 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.848329 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.848341 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.951089 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.951126 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.951137 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.951154 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:56 crc kubenswrapper[4876]: I0313 07:40:56.951164 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:56Z","lastTransitionTime":"2026-03-13T07:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.053761 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.054058 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.054164 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.054279 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.054374 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.156970 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.157007 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.157018 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.157033 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.157044 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.260230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.260291 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.260302 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.260322 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.260333 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.363559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.363874 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.363975 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.364081 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.364170 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.467186 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.467510 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.467642 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.467773 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.467929 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.487773 4876 generic.go:334] "Generic (PLEG): container finished" podID="153920eb-fdb1-4e94-85ca-f09b5f01c2aa" containerID="e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3" exitCode=0 Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.487908 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" event={"ID":"153920eb-fdb1-4e94-85ca-f09b5f01c2aa","Type":"ContainerDied","Data":"e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.493641 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.493691 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.495517 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.517489 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.532169 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.541747 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.552537 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.567752 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.571546 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.571585 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.571598 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.571614 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.571625 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.582337 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.603796 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.617956 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.630554 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.645853 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.658764 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.673745 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.673779 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.673787 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.673802 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.673811 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.676942 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.688576 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.699266 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.709558 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.722456 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.737463 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.746291 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.757224 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.769133 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.776661 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.776685 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.776694 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.776707 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.776717 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.782367 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.782434 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.782463 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.782483 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.782501 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782562 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782594 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:01.782584217 +0000 UTC m=+121.453363199 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782781 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:41:01.782773422 +0000 UTC m=+121.453552404 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782835 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782856 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:01.782850784 +0000 UTC m=+121.453629766 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782899 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782910 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782919 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782939 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:01.782933306 +0000 UTC m=+121.453712288 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782974 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782983 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.782992 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.783009 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:01.783003849 +0000 UTC m=+121.453782831 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.786292 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.797636 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.807860 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.822182 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.837179 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.848219 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.860208 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.871457 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.879014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.879050 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.879060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.879073 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.879083 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.880761 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.882979 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.883091 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: E0313 07:40:57.883155 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs podName:f36294fb-98b9-48be-8237-beb6484e0cb6 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:01.883137582 +0000 UTC m=+121.553916574 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs") pod "network-metrics-daemon-zrbh4" (UID: "f36294fb-98b9-48be-8237-beb6484e0cb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.890276 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:57Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.981350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.981597 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.981606 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.981618 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:57 crc kubenswrapper[4876]: I0313 07:40:57.981627 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:57Z","lastTransitionTime":"2026-03-13T07:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.035354 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.035423 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.035435 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:40:58 crc kubenswrapper[4876]: E0313 07:40:58.035566 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.035955 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:40:58 crc kubenswrapper[4876]: E0313 07:40:58.036036 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:40:58 crc kubenswrapper[4876]: E0313 07:40:58.036102 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:40:58 crc kubenswrapper[4876]: E0313 07:40:58.036155 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.086152 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.086249 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.086338 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.086356 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.086367 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.188304 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.188328 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.188336 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.188347 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.188356 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.290324 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.290357 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.290367 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.290380 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.290389 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.392933 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.393039 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.393062 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.393087 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.393104 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.495145 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.495175 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.495186 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.495200 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.495211 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.500157 4876 generic.go:334] "Generic (PLEG): container finished" podID="153920eb-fdb1-4e94-85ca-f09b5f01c2aa" containerID="2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73" exitCode=0 Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.500219 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" event={"ID":"153920eb-fdb1-4e94-85ca-f09b5f01c2aa","Type":"ContainerDied","Data":"2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.512786 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.543523 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.559265 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.571785 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.584869 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.596149 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.600636 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.600664 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.600672 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.600687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.600696 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.607928 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.617912 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.627377 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.636091 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.647227 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.656539 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.670694 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.682853 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.693985 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:58Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.702759 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.702787 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.702795 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.702808 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.702817 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.805516 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.805550 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.805561 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.805575 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.805585 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.908126 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.908168 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.908185 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.908200 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:58 crc kubenswrapper[4876]: I0313 07:40:58.908210 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:58Z","lastTransitionTime":"2026-03-13T07:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.017651 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.017704 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.017717 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.017732 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.018047 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.120283 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.120313 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.120321 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.120333 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.120343 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.224205 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.224311 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.224334 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.224361 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.224379 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.327668 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.327706 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.327714 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.327727 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.327737 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.429780 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.429814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.429824 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.429838 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.429847 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.508059 4876 generic.go:334] "Generic (PLEG): container finished" podID="153920eb-fdb1-4e94-85ca-f09b5f01c2aa" containerID="d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0" exitCode=0 Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.508188 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" event={"ID":"153920eb-fdb1-4e94-85ca-f09b5f01c2aa","Type":"ContainerDied","Data":"d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.516513 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.529006 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.534203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.534259 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.534271 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.534291 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.534304 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.545481 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.559261 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.572901 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.586277 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.597801 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.612487 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.628532 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.636765 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.636810 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.636823 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.636840 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.636852 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.639650 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.652904 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.667655 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.682019 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.695681 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.695727 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.695740 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.695755 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.695765 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.703472 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: E0313 07:40:59.713706 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.717332 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.717370 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.717381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.717398 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.717407 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.719644 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: E0313 07:40:59.729659 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.731409 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.733111 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.733149 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.733180 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.733197 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.733210 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: E0313 07:40:59.745083 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.748566 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.748711 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.748779 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.748848 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.748911 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: E0313 07:40:59.761845 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.765833 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.765882 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.765890 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.765904 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.765915 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: E0313 07:40:59.777864 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:40:59Z is after 2025-08-24T17:21:41Z" Mar 13 07:40:59 crc kubenswrapper[4876]: E0313 07:40:59.778041 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.779517 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.779549 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.779561 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.779577 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.779588 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.882691 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.882735 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.882745 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.882761 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.882771 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.985333 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.985396 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.985410 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.985427 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:40:59 crc kubenswrapper[4876]: I0313 07:40:59.985440 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:40:59Z","lastTransitionTime":"2026-03-13T07:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.034492 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.034521 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.034568 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.034637 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:00 crc kubenswrapper[4876]: E0313 07:41:00.034832 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:00 crc kubenswrapper[4876]: E0313 07:41:00.035427 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:00 crc kubenswrapper[4876]: E0313 07:41:00.035500 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:00 crc kubenswrapper[4876]: E0313 07:41:00.035690 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.087877 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.087928 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.087944 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.087965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.087981 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.191232 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.191303 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.191319 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.191337 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.191347 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.293688 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.293727 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.293737 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.293752 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.293763 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.396315 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.396656 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.396686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.396714 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.396735 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.498873 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.498914 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.498926 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.498943 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.498954 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.524592 4876 generic.go:334] "Generic (PLEG): container finished" podID="153920eb-fdb1-4e94-85ca-f09b5f01c2aa" containerID="f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04" exitCode=0 Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.524643 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" event={"ID":"153920eb-fdb1-4e94-85ca-f09b5f01c2aa","Type":"ContainerDied","Data":"f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.537956 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.550889 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.561484 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.575928 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.588856 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.600711 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.600768 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.600780 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.600796 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.600807 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.605504 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.619313 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.629776 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.641099 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.655302 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.669772 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.695026 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.703436 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.703471 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.703479 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.703493 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.703502 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.721498 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.733696 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.751494 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:00Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.805791 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.805829 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.805839 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.805855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.805867 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.907433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.907469 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.907477 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.907491 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:00 crc kubenswrapper[4876]: I0313 07:41:00.907502 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:00Z","lastTransitionTime":"2026-03-13T07:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.008020 4876 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.051699 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.064590 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.081265 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.095084 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.106001 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.116330 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.129493 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.146696 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.164770 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.166138 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.183258 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.194725 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.205166 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.215396 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.229447 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.242567 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.534422 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb"} Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.534965 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.535010 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.542524 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" event={"ID":"153920eb-fdb1-4e94-85ca-f09b5f01c2aa","Type":"ContainerStarted","Data":"fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f"} Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.554169 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.568606 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.571951 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.584314 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.604472 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.625840 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.642694 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.663474 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.678331 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.692321 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.707107 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.721401 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.737779 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.750950 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.762901 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.772870 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.784443 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.799889 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.811019 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.823365 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.832322 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.832497 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832525 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:41:09.832502246 +0000 UTC m=+129.503281228 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.832561 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.832631 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832653 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832680 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.832686 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832696 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832754 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:09.832734462 +0000 UTC m=+129.503513524 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832794 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832815 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832836 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832846 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832875 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:09.832856775 +0000 UTC m=+129.503635757 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832894 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:09.832887346 +0000 UTC m=+129.503666328 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832915 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.832946 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:09.832937707 +0000 UTC m=+129.503716689 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.840489 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.863857 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.877607 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.893013 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.906898 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.922732 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.933488 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.933666 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: E0313 07:41:01.933738 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs podName:f36294fb-98b9-48be-8237-beb6484e0cb6 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:09.933720659 +0000 UTC m=+129.604499641 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs") pod "network-metrics-daemon-zrbh4" (UID: "f36294fb-98b9-48be-8237-beb6484e0cb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.936485 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.945965 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.957556 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.970887 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:01 crc kubenswrapper[4876]: I0313 07:41:01.981693 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.034862 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.034963 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:02 crc kubenswrapper[4876]: E0313 07:41:02.035053 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.035131 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.034876 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:02 crc kubenswrapper[4876]: E0313 07:41:02.035403 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:02 crc kubenswrapper[4876]: E0313 07:41:02.035347 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:02 crc kubenswrapper[4876]: E0313 07:41:02.035716 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.546215 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.566122 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.583614 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.596912 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.607871 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.620861 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.638752 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.659039 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.672261 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.688404 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.704515 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.732538 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.754514 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.773034 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.789405 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.806992 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:02 crc kubenswrapper[4876]: I0313 07:41:02.827273 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:02Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:03 crc kubenswrapper[4876]: I0313 07:41:03.051618 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.034893 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.034984 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.035008 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:04 crc kubenswrapper[4876]: E0313 07:41:04.035547 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:04 crc kubenswrapper[4876]: E0313 07:41:04.035354 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.035010 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:04 crc kubenswrapper[4876]: E0313 07:41:04.035627 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:04 crc kubenswrapper[4876]: E0313 07:41:04.035673 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.553189 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/0.log" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.555540 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb" exitCode=1 Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.555578 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb"} Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.556160 4876 scope.go:117] "RemoveContainer" containerID="df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.576462 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.594777 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.607510 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.622123 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.636815 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.648967 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.668376 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:03Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:140\\\\nI0313 07:41:03.784716 6796 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:03.784749 6796 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 07:41:03.784776 6796 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 07:41:03.784830 6796 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:03.784862 6796 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 07:41:03.784968 6796 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 07:41:03.785308 6796 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 07:41:03.785728 6796 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.684405 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.696515 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.708554 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.727707 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.739623 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.750932 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.761324 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.771824 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.783830 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:04Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:04 crc kubenswrapper[4876]: I0313 07:41:04.952183 4876 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.560747 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/1.log" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.561169 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/0.log" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.564476 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d" exitCode=1 Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.564511 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d"} Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.564550 4876 scope.go:117] "RemoveContainer" containerID="df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.565601 4876 scope.go:117] "RemoveContainer" containerID="a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d" Mar 13 07:41:05 crc kubenswrapper[4876]: E0313 07:41:05.565866 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.580336 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.600278 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.614675 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.628092 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.638886 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.651825 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.664889 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.684723 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.707614 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df728b24212013cdffda5e33fc314e707e126a624e3fc473f9245c48afd020cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:03Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:140\\\\nI0313 07:41:03.784716 6796 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:03.784749 6796 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 07:41:03.784776 6796 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 07:41:03.784830 6796 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:03.784862 6796 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 07:41:03.784968 6796 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 07:41:03.785308 6796 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 07:41:03.785728 6796 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:05Z\\\",\\\"message\\\":\\\"60d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 07:41:05.340598 6932 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}\\\\nI0313 07:41:05.340704 6932 services_controller.go:360] Finished syncing service kube-controller-manager on namespace openshift-kube-controller-manager for network=default : 2.243261ms\\\\nI0313 07:41:05.340738 6932 services_controller.go:356] Processing sync for service openshift-ovn-kubernetes/ovn-kubernetes-node for network=default\\\\nI0313 07:41:05.340745 6932 services_controller.go:360] Finished syncing service ovn-kubernetes-node on namespace openshift-ovn-kubernetes for network=default : 8.25µs\\\\nI0313 07:41:05.340753 6932 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nI0313 07:41:05.339941 6932 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0313 07:41:05.340789 6932 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0313 07:41:05.340776 6932 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.723204 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.737101 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.750567 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.764384 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.786795 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.802623 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:05 crc kubenswrapper[4876]: I0313 07:41:05.820701 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:05Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.034903 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.034950 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.035031 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:06 crc kubenswrapper[4876]: E0313 07:41:06.035200 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:06 crc kubenswrapper[4876]: E0313 07:41:06.035326 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:06 crc kubenswrapper[4876]: E0313 07:41:06.035412 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.035824 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:06 crc kubenswrapper[4876]: E0313 07:41:06.035885 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:06 crc kubenswrapper[4876]: E0313 07:41:06.167754 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.570774 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/1.log" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.576672 4876 scope.go:117] "RemoveContainer" containerID="a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d" Mar 13 07:41:06 crc kubenswrapper[4876]: E0313 07:41:06.577134 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.592928 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.612570 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.633165 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.648814 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.665470 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.680157 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.698158 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.719918 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.733735 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.750135 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:05Z\\\",\\\"message\\\":\\\"60d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 07:41:05.340598 6932 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}\\\\nI0313 07:41:05.340704 6932 services_controller.go:360] Finished syncing service kube-controller-manager on namespace openshift-kube-controller-manager for network=default : 2.243261ms\\\\nI0313 07:41:05.340738 6932 services_controller.go:356] Processing sync for service openshift-ovn-kubernetes/ovn-kubernetes-node for network=default\\\\nI0313 07:41:05.340745 6932 services_controller.go:360] Finished syncing service ovn-kubernetes-node on namespace openshift-ovn-kubernetes for network=default : 8.25µs\\\\nI0313 07:41:05.340753 6932 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nI0313 07:41:05.339941 6932 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0313 07:41:05.340789 6932 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0313 07:41:05.340776 6932 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.761621 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.779757 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.795929 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.807195 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.817919 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:06 crc kubenswrapper[4876]: I0313 07:41:06.828290 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:06Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:08 crc kubenswrapper[4876]: I0313 07:41:08.035109 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:08 crc kubenswrapper[4876]: I0313 07:41:08.035154 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:08 crc kubenswrapper[4876]: I0313 07:41:08.035152 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:08 crc kubenswrapper[4876]: I0313 07:41:08.035152 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:08 crc kubenswrapper[4876]: E0313 07:41:08.035298 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:08 crc kubenswrapper[4876]: E0313 07:41:08.035405 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:08 crc kubenswrapper[4876]: E0313 07:41:08.035621 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:08 crc kubenswrapper[4876]: E0313 07:41:08.035665 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:08 crc kubenswrapper[4876]: I0313 07:41:08.045389 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 13 07:41:09 crc kubenswrapper[4876]: I0313 07:41:09.918108 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918330 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:41:25.918310958 +0000 UTC m=+145.589089930 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:41:09 crc kubenswrapper[4876]: I0313 07:41:09.918594 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:09 crc kubenswrapper[4876]: I0313 07:41:09.918628 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:09 crc kubenswrapper[4876]: I0313 07:41:09.918653 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:09 crc kubenswrapper[4876]: I0313 07:41:09.918684 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918773 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918782 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918792 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918805 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918819 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918840 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:25.918828882 +0000 UTC m=+145.589607864 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918856 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:25.918849192 +0000 UTC m=+145.589628174 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918869 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:25.918863093 +0000 UTC m=+145.589642075 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918879 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918928 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.918941 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:09 crc kubenswrapper[4876]: E0313 07:41:09.919010 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:25.918991366 +0000 UTC m=+145.589770468 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.019897 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.020089 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.020189 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs podName:f36294fb-98b9-48be-8237-beb6484e0cb6 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:26.020168858 +0000 UTC m=+145.690947890 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs") pod "network-metrics-daemon-zrbh4" (UID: "f36294fb-98b9-48be-8237-beb6484e0cb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.035312 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.035425 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.035469 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.035601 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.035627 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.035708 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.035724 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.035883 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.089542 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.089618 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.089640 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.089817 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.089839 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:10Z","lastTransitionTime":"2026-03-13T07:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.102871 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:10Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.106636 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.106700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.106720 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.106744 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.106762 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:10Z","lastTransitionTime":"2026-03-13T07:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.120986 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:10Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.124752 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.124793 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.124805 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.124828 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.124841 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:10Z","lastTransitionTime":"2026-03-13T07:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.137014 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:10Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.140433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.140457 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.140465 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.140478 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.140487 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:10Z","lastTransitionTime":"2026-03-13T07:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.151329 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:10Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.156164 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.156217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.156260 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.156277 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:10 crc kubenswrapper[4876]: I0313 07:41:10.156348 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:10Z","lastTransitionTime":"2026-03-13T07:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.167155 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:10Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:10 crc kubenswrapper[4876]: E0313 07:41:10.167313 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.048396 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.063427 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.085142 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.100552 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.115228 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.133310 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:05Z\\\",\\\"message\\\":\\\"60d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 07:41:05.340598 6932 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}\\\\nI0313 07:41:05.340704 6932 services_controller.go:360] Finished syncing service kube-controller-manager on namespace openshift-kube-controller-manager for network=default : 2.243261ms\\\\nI0313 07:41:05.340738 6932 services_controller.go:356] Processing sync for service openshift-ovn-kubernetes/ovn-kubernetes-node for network=default\\\\nI0313 07:41:05.340745 6932 services_controller.go:360] Finished syncing service ovn-kubernetes-node on namespace openshift-ovn-kubernetes for network=default : 8.25µs\\\\nI0313 07:41:05.340753 6932 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nI0313 07:41:05.339941 6932 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0313 07:41:05.340789 6932 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0313 07:41:05.340776 6932 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.145088 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.159879 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: E0313 07:41:11.168984 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.178978 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.191976 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.209547 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.223442 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.235636 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.250600 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.267439 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.284559 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:11 crc kubenswrapper[4876]: I0313 07:41:11.297938 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:11Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:12 crc kubenswrapper[4876]: I0313 07:41:12.034510 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:12 crc kubenswrapper[4876]: I0313 07:41:12.034552 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:12 crc kubenswrapper[4876]: E0313 07:41:12.034714 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:12 crc kubenswrapper[4876]: I0313 07:41:12.034510 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:12 crc kubenswrapper[4876]: E0313 07:41:12.034906 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:12 crc kubenswrapper[4876]: I0313 07:41:12.034962 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:12 crc kubenswrapper[4876]: E0313 07:41:12.035042 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:12 crc kubenswrapper[4876]: E0313 07:41:12.035173 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.552619 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.567447 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.586682 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.605930 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:05Z\\\",\\\"message\\\":\\\"60d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 07:41:05.340598 6932 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}\\\\nI0313 07:41:05.340704 6932 services_controller.go:360] Finished syncing service kube-controller-manager on namespace openshift-kube-controller-manager for network=default : 2.243261ms\\\\nI0313 07:41:05.340738 6932 services_controller.go:356] Processing sync for service openshift-ovn-kubernetes/ovn-kubernetes-node for network=default\\\\nI0313 07:41:05.340745 6932 services_controller.go:360] Finished syncing service ovn-kubernetes-node on namespace openshift-ovn-kubernetes for network=default : 8.25µs\\\\nI0313 07:41:05.340753 6932 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nI0313 07:41:05.339941 6932 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0313 07:41:05.340789 6932 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0313 07:41:05.340776 6932 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.620057 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.634056 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.646936 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.662529 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.675776 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.686740 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.697149 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.705699 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.714317 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.724573 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.737096 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.751894 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.762306 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:13 crc kubenswrapper[4876]: I0313 07:41:13.773014 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:13Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:14 crc kubenswrapper[4876]: I0313 07:41:14.034919 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:14 crc kubenswrapper[4876]: I0313 07:41:14.034977 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:14 crc kubenswrapper[4876]: I0313 07:41:14.034979 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:14 crc kubenswrapper[4876]: I0313 07:41:14.034991 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:14 crc kubenswrapper[4876]: E0313 07:41:14.035083 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:14 crc kubenswrapper[4876]: E0313 07:41:14.035218 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:14 crc kubenswrapper[4876]: E0313 07:41:14.035324 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:14 crc kubenswrapper[4876]: E0313 07:41:14.035434 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:16 crc kubenswrapper[4876]: I0313 07:41:16.035028 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:16 crc kubenswrapper[4876]: E0313 07:41:16.035180 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:16 crc kubenswrapper[4876]: I0313 07:41:16.035277 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:16 crc kubenswrapper[4876]: I0313 07:41:16.035371 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:16 crc kubenswrapper[4876]: I0313 07:41:16.035507 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:16 crc kubenswrapper[4876]: E0313 07:41:16.035507 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:16 crc kubenswrapper[4876]: E0313 07:41:16.035582 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:16 crc kubenswrapper[4876]: E0313 07:41:16.035647 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:16 crc kubenswrapper[4876]: E0313 07:41:16.171371 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:18 crc kubenswrapper[4876]: I0313 07:41:18.035338 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:18 crc kubenswrapper[4876]: I0313 07:41:18.035390 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:18 crc kubenswrapper[4876]: I0313 07:41:18.035432 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:18 crc kubenswrapper[4876]: I0313 07:41:18.035360 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:18 crc kubenswrapper[4876]: E0313 07:41:18.035534 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:18 crc kubenswrapper[4876]: E0313 07:41:18.035654 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:18 crc kubenswrapper[4876]: E0313 07:41:18.035753 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:18 crc kubenswrapper[4876]: E0313 07:41:18.035855 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.035114 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.035217 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.035286 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.035349 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.035431 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.035478 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.035534 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.035662 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.036578 4876 scope.go:117] "RemoveContainer" containerID="a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.429846 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.430106 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.430115 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.430127 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.430136 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:20Z","lastTransitionTime":"2026-03-13T07:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.441419 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.444207 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.444253 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.444264 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.444283 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.444294 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:20Z","lastTransitionTime":"2026-03-13T07:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.454377 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.457118 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.457147 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.457158 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.457171 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.457181 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:20Z","lastTransitionTime":"2026-03-13T07:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.467052 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.470047 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.470107 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.470116 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.470128 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.470137 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:20Z","lastTransitionTime":"2026-03-13T07:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.479951 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.482690 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.482769 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.482783 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.482795 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.482805 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:20Z","lastTransitionTime":"2026-03-13T07:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.492794 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: E0313 07:41:20.492904 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.623436 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/1.log" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.625620 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4"} Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.626077 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.637841 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.646576 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.655953 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.666649 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.677780 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.692655 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.703344 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.716862 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.729472 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.744105 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.758344 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.777989 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:05Z\\\",\\\"message\\\":\\\"60d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 07:41:05.340598 6932 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}\\\\nI0313 07:41:05.340704 6932 services_controller.go:360] Finished syncing service kube-controller-manager on namespace openshift-kube-controller-manager for network=default : 2.243261ms\\\\nI0313 07:41:05.340738 6932 services_controller.go:356] Processing sync for service openshift-ovn-kubernetes/ovn-kubernetes-node for network=default\\\\nI0313 07:41:05.340745 6932 services_controller.go:360] Finished syncing service ovn-kubernetes-node on namespace openshift-ovn-kubernetes for network=default : 8.25µs\\\\nI0313 07:41:05.340753 6932 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nI0313 07:41:05.339941 6932 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0313 07:41:05.340789 6932 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0313 07:41:05.340776 6932 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.790225 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.799500 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.810768 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.820707 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:20 crc kubenswrapper[4876]: I0313 07:41:20.830014 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:20Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.050927 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.065215 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.077845 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.090404 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.113044 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.129215 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.140740 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.162378 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:05Z\\\",\\\"message\\\":\\\"60d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 07:41:05.340598 6932 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}\\\\nI0313 07:41:05.340704 6932 services_controller.go:360] Finished syncing service kube-controller-manager on namespace openshift-kube-controller-manager for network=default : 2.243261ms\\\\nI0313 07:41:05.340738 6932 services_controller.go:356] Processing sync for service openshift-ovn-kubernetes/ovn-kubernetes-node for network=default\\\\nI0313 07:41:05.340745 6932 services_controller.go:360] Finished syncing service ovn-kubernetes-node on namespace openshift-ovn-kubernetes for network=default : 8.25µs\\\\nI0313 07:41:05.340753 6932 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nI0313 07:41:05.339941 6932 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0313 07:41:05.340789 6932 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0313 07:41:05.340776 6932 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: E0313 07:41:21.171688 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.175929 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.184771 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.194558 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.208886 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.219029 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.230528 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.242480 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.250717 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.258794 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.631926 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/2.log" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.632585 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/1.log" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.636412 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4" exitCode=1 Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.636499 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4"} Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.636556 4876 scope.go:117] "RemoveContainer" containerID="a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.637181 4876 scope.go:117] "RemoveContainer" containerID="29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4" Mar 13 07:41:21 crc kubenswrapper[4876]: E0313 07:41:21.637416 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.656908 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.674109 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.688974 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.708806 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.725183 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.743171 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.755715 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.769605 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.784131 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.799532 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.818205 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a88d70ae2733146a44c139f392b21bd40baf2f9a27db6a668ca1855125556d2d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:05Z\\\",\\\"message\\\":\\\"60d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 07:41:05.340598 6932 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}\\\\nI0313 07:41:05.340704 6932 services_controller.go:360] Finished syncing service kube-controller-manager on namespace openshift-kube-controller-manager for network=default : 2.243261ms\\\\nI0313 07:41:05.340738 6932 services_controller.go:356] Processing sync for service openshift-ovn-kubernetes/ovn-kubernetes-node for network=default\\\\nI0313 07:41:05.340745 6932 services_controller.go:360] Finished syncing service ovn-kubernetes-node on namespace openshift-ovn-kubernetes for network=default : 8.25µs\\\\nI0313 07:41:05.340753 6932 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nI0313 07:41:05.339941 6932 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI0313 07:41:05.340789 6932 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI0313 07:41:05.340776 6932 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.829874 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.841701 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.857002 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.876941 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.890284 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:21 crc kubenswrapper[4876]: I0313 07:41:21.901620 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:21Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.035020 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.035109 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:22 crc kubenswrapper[4876]: E0313 07:41:22.035147 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.035019 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.035035 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:22 crc kubenswrapper[4876]: E0313 07:41:22.035279 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:22 crc kubenswrapper[4876]: E0313 07:41:22.035319 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:22 crc kubenswrapper[4876]: E0313 07:41:22.035375 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.640886 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/2.log" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.645318 4876 scope.go:117] "RemoveContainer" containerID="29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4" Mar 13 07:41:22 crc kubenswrapper[4876]: E0313 07:41:22.645729 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.657618 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.668825 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.680874 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.700335 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.722482 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.736685 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.752900 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.770720 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.790216 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.805427 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.826018 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.848260 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.859586 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.876147 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.888132 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.900198 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:22 crc kubenswrapper[4876]: I0313 07:41:22.911075 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:22Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:24 crc kubenswrapper[4876]: I0313 07:41:24.034915 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:24 crc kubenswrapper[4876]: I0313 07:41:24.034935 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:24 crc kubenswrapper[4876]: I0313 07:41:24.034961 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:24 crc kubenswrapper[4876]: I0313 07:41:24.035108 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:24 crc kubenswrapper[4876]: E0313 07:41:24.035386 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:24 crc kubenswrapper[4876]: E0313 07:41:24.035649 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:24 crc kubenswrapper[4876]: E0313 07:41:24.035734 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:24 crc kubenswrapper[4876]: E0313 07:41:24.035839 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:25 crc kubenswrapper[4876]: I0313 07:41:25.972902 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:41:25 crc kubenswrapper[4876]: I0313 07:41:25.973045 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973078 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:41:57.973057416 +0000 UTC m=+177.643836388 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:41:25 crc kubenswrapper[4876]: I0313 07:41:25.973110 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:25 crc kubenswrapper[4876]: I0313 07:41:25.973144 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:25 crc kubenswrapper[4876]: I0313 07:41:25.973186 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973148 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973308 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973313 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973347 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973173 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973377 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973358 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973419 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973347 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:57.973338363 +0000 UTC m=+177.644117345 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973479 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:57.973455857 +0000 UTC m=+177.644234899 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973498 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:57.973490757 +0000 UTC m=+177.644269849 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:25 crc kubenswrapper[4876]: E0313 07:41:25.973514 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:57.973506208 +0000 UTC m=+177.644285320 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:26 crc kubenswrapper[4876]: I0313 07:41:26.034852 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:26 crc kubenswrapper[4876]: I0313 07:41:26.034865 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:26 crc kubenswrapper[4876]: I0313 07:41:26.034909 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:26 crc kubenswrapper[4876]: I0313 07:41:26.034961 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:26 crc kubenswrapper[4876]: E0313 07:41:26.034979 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:26 crc kubenswrapper[4876]: E0313 07:41:26.035092 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:26 crc kubenswrapper[4876]: E0313 07:41:26.035181 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:26 crc kubenswrapper[4876]: E0313 07:41:26.035298 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:26 crc kubenswrapper[4876]: I0313 07:41:26.074060 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:26 crc kubenswrapper[4876]: E0313 07:41:26.074322 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:41:26 crc kubenswrapper[4876]: E0313 07:41:26.074418 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs podName:f36294fb-98b9-48be-8237-beb6484e0cb6 nodeName:}" failed. No retries permitted until 2026-03-13 07:41:58.074401748 +0000 UTC m=+177.745180730 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs") pod "network-metrics-daemon-zrbh4" (UID: "f36294fb-98b9-48be-8237-beb6484e0cb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:41:26 crc kubenswrapper[4876]: E0313 07:41:26.173215 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:28 crc kubenswrapper[4876]: I0313 07:41:28.035219 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:28 crc kubenswrapper[4876]: I0313 07:41:28.035293 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:28 crc kubenswrapper[4876]: I0313 07:41:28.035406 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:28 crc kubenswrapper[4876]: E0313 07:41:28.035519 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:28 crc kubenswrapper[4876]: I0313 07:41:28.035550 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:28 crc kubenswrapper[4876]: E0313 07:41:28.035663 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:28 crc kubenswrapper[4876]: E0313 07:41:28.035763 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:28 crc kubenswrapper[4876]: E0313 07:41:28.035825 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.035414 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.035862 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.035481 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.035935 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.035489 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.035983 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.035431 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.036024 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.830345 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.830391 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.830401 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.830418 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.830429 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:30Z","lastTransitionTime":"2026-03-13T07:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.844144 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:30Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.848445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.848497 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.848514 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.848541 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.848564 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:30Z","lastTransitionTime":"2026-03-13T07:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.863369 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:30Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.866559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.866604 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.866615 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.866633 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.866645 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:30Z","lastTransitionTime":"2026-03-13T07:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.880050 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:30Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.883730 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.883765 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.883774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.883788 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.883797 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:30Z","lastTransitionTime":"2026-03-13T07:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.898551 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:30Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.902500 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.902536 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.902545 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.902559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:30 crc kubenswrapper[4876]: I0313 07:41:30.902568 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:30Z","lastTransitionTime":"2026-03-13T07:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.915717 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:30Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:30 crc kubenswrapper[4876]: E0313 07:41:30.915866 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.052297 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.062138 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.072297 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.083981 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.096668 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.110062 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.122276 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.134590 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.150317 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.163041 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: E0313 07:41:31.173769 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.182323 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.194590 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.205830 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.215924 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.228819 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.239377 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:31 crc kubenswrapper[4876]: I0313 07:41:31.249735 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:31Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:32 crc kubenswrapper[4876]: I0313 07:41:32.035507 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:32 crc kubenswrapper[4876]: I0313 07:41:32.035679 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:32 crc kubenswrapper[4876]: I0313 07:41:32.035679 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:32 crc kubenswrapper[4876]: I0313 07:41:32.035843 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:32 crc kubenswrapper[4876]: E0313 07:41:32.035845 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:32 crc kubenswrapper[4876]: E0313 07:41:32.035920 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:32 crc kubenswrapper[4876]: E0313 07:41:32.035995 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:32 crc kubenswrapper[4876]: E0313 07:41:32.036049 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:34 crc kubenswrapper[4876]: I0313 07:41:34.035317 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:34 crc kubenswrapper[4876]: I0313 07:41:34.035390 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:34 crc kubenswrapper[4876]: I0313 07:41:34.035415 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:34 crc kubenswrapper[4876]: I0313 07:41:34.035341 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:34 crc kubenswrapper[4876]: E0313 07:41:34.035523 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:34 crc kubenswrapper[4876]: E0313 07:41:34.035676 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:34 crc kubenswrapper[4876]: E0313 07:41:34.035802 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:34 crc kubenswrapper[4876]: E0313 07:41:34.035931 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:35 crc kubenswrapper[4876]: I0313 07:41:35.036505 4876 scope.go:117] "RemoveContainer" containerID="29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4" Mar 13 07:41:35 crc kubenswrapper[4876]: E0313 07:41:35.037028 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:41:36 crc kubenswrapper[4876]: I0313 07:41:36.034945 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:36 crc kubenswrapper[4876]: I0313 07:41:36.035018 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:36 crc kubenswrapper[4876]: I0313 07:41:36.035048 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:36 crc kubenswrapper[4876]: E0313 07:41:36.035170 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:36 crc kubenswrapper[4876]: E0313 07:41:36.035386 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:36 crc kubenswrapper[4876]: E0313 07:41:36.035541 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:36 crc kubenswrapper[4876]: I0313 07:41:36.035573 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:36 crc kubenswrapper[4876]: E0313 07:41:36.035744 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:36 crc kubenswrapper[4876]: E0313 07:41:36.176000 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:38 crc kubenswrapper[4876]: I0313 07:41:38.035409 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:38 crc kubenswrapper[4876]: E0313 07:41:38.035564 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:38 crc kubenswrapper[4876]: I0313 07:41:38.035435 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:38 crc kubenswrapper[4876]: I0313 07:41:38.035668 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:38 crc kubenswrapper[4876]: I0313 07:41:38.035420 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:38 crc kubenswrapper[4876]: E0313 07:41:38.035869 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:38 crc kubenswrapper[4876]: E0313 07:41:38.035923 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:38 crc kubenswrapper[4876]: E0313 07:41:38.036051 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:40 crc kubenswrapper[4876]: I0313 07:41:40.035319 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:40 crc kubenswrapper[4876]: I0313 07:41:40.035362 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:40 crc kubenswrapper[4876]: I0313 07:41:40.035359 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:40 crc kubenswrapper[4876]: E0313 07:41:40.035458 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:40 crc kubenswrapper[4876]: I0313 07:41:40.035494 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:40 crc kubenswrapper[4876]: E0313 07:41:40.035773 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:40 crc kubenswrapper[4876]: E0313 07:41:40.035779 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:40 crc kubenswrapper[4876]: E0313 07:41:40.036009 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.048041 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.060705 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.074004 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.085210 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.097346 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.113527 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.132997 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.158951 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.175333 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: E0313 07:41:41.176847 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.187357 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.198048 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.211453 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.231217 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.241857 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.253167 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.262738 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.273548 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.292036 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.292081 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.292093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.292109 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.292121 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:41Z","lastTransitionTime":"2026-03-13T07:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:41 crc kubenswrapper[4876]: E0313 07:41:41.303824 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.307969 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.308003 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.308011 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.308025 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.308034 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:41Z","lastTransitionTime":"2026-03-13T07:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:41 crc kubenswrapper[4876]: E0313 07:41:41.319888 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.324090 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.324123 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.324134 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.324151 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.324162 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:41Z","lastTransitionTime":"2026-03-13T07:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:41 crc kubenswrapper[4876]: E0313 07:41:41.336776 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.340148 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.340191 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.340201 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.340219 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.340230 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:41Z","lastTransitionTime":"2026-03-13T07:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:41 crc kubenswrapper[4876]: E0313 07:41:41.353521 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.357219 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.357284 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.357301 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.357325 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.357341 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:41Z","lastTransitionTime":"2026-03-13T07:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:41 crc kubenswrapper[4876]: E0313 07:41:41.370281 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: E0313 07:41:41.370403 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.715166 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/0.log" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.715220 4876 generic.go:334] "Generic (PLEG): container finished" podID="8adb2a69-d0ef-4efc-813d-77fdf9713d50" containerID="77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369" exitCode=1 Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.715268 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-st667" event={"ID":"8adb2a69-d0ef-4efc-813d-77fdf9713d50","Type":"ContainerDied","Data":"77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369"} Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.716342 4876 scope.go:117] "RemoveContainer" containerID="77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.728161 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.747002 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.759365 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.772755 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.787682 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.808011 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"2026-03-13T07:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d\\\\n2026-03-13T07:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d to /host/opt/cni/bin/\\\\n2026-03-13T07:40:56Z [verbose] multus-daemon started\\\\n2026-03-13T07:40:56Z [verbose] Readiness Indicator file check\\\\n2026-03-13T07:41:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.820658 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.834019 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.847142 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.861858 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.874266 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.888680 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.902956 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.925598 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.940719 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.957963 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:41 crc kubenswrapper[4876]: I0313 07:41:41.975972 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:41Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.034465 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.034527 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.034480 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:42 crc kubenswrapper[4876]: E0313 07:41:42.034651 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.034696 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:42 crc kubenswrapper[4876]: E0313 07:41:42.034798 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:42 crc kubenswrapper[4876]: E0313 07:41:42.034971 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:42 crc kubenswrapper[4876]: E0313 07:41:42.035197 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.723313 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/0.log" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.723404 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-st667" event={"ID":"8adb2a69-d0ef-4efc-813d-77fdf9713d50","Type":"ContainerStarted","Data":"a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4"} Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.745353 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.765955 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.783914 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.812385 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.828393 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.844080 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.863612 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.886789 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.902796 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.920743 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.937572 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.953405 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.966118 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:42 crc kubenswrapper[4876]: I0313 07:41:42.983576 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:43 crc kubenswrapper[4876]: I0313 07:41:43.002306 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:42Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:43 crc kubenswrapper[4876]: I0313 07:41:43.022526 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"2026-03-13T07:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d\\\\n2026-03-13T07:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d to /host/opt/cni/bin/\\\\n2026-03-13T07:40:56Z [verbose] multus-daemon started\\\\n2026-03-13T07:40:56Z [verbose] Readiness Indicator file check\\\\n2026-03-13T07:41:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:43Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:43 crc kubenswrapper[4876]: I0313 07:41:43.037556 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:43Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:44 crc kubenswrapper[4876]: I0313 07:41:44.035673 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:44 crc kubenswrapper[4876]: I0313 07:41:44.035759 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:44 crc kubenswrapper[4876]: I0313 07:41:44.035795 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:44 crc kubenswrapper[4876]: E0313 07:41:44.035876 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:44 crc kubenswrapper[4876]: E0313 07:41:44.036073 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:44 crc kubenswrapper[4876]: I0313 07:41:44.036165 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:44 crc kubenswrapper[4876]: E0313 07:41:44.036406 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:44 crc kubenswrapper[4876]: E0313 07:41:44.036547 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:44 crc kubenswrapper[4876]: I0313 07:41:44.056775 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 13 07:41:46 crc kubenswrapper[4876]: I0313 07:41:46.034934 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:46 crc kubenswrapper[4876]: I0313 07:41:46.034958 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:46 crc kubenswrapper[4876]: I0313 07:41:46.035032 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:46 crc kubenswrapper[4876]: I0313 07:41:46.034935 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:46 crc kubenswrapper[4876]: E0313 07:41:46.035135 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:46 crc kubenswrapper[4876]: E0313 07:41:46.035067 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:46 crc kubenswrapper[4876]: E0313 07:41:46.035291 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:46 crc kubenswrapper[4876]: E0313 07:41:46.035403 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:46 crc kubenswrapper[4876]: E0313 07:41:46.178828 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:48 crc kubenswrapper[4876]: I0313 07:41:48.034495 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:48 crc kubenswrapper[4876]: I0313 07:41:48.034560 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:48 crc kubenswrapper[4876]: E0313 07:41:48.034623 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:48 crc kubenswrapper[4876]: I0313 07:41:48.034495 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:48 crc kubenswrapper[4876]: I0313 07:41:48.034512 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:48 crc kubenswrapper[4876]: E0313 07:41:48.034768 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:48 crc kubenswrapper[4876]: E0313 07:41:48.034695 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:48 crc kubenswrapper[4876]: E0313 07:41:48.034919 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.035314 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:50 crc kubenswrapper[4876]: E0313 07:41:50.035659 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.035715 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.035815 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:50 crc kubenswrapper[4876]: E0313 07:41:50.036563 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.036624 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.036994 4876 scope.go:117] "RemoveContainer" containerID="29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4" Mar 13 07:41:50 crc kubenswrapper[4876]: E0313 07:41:50.037065 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:50 crc kubenswrapper[4876]: E0313 07:41:50.037121 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.048923 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.759340 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/2.log" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.761673 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.775132 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.786233 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"2026-03-13T07:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d\\\\n2026-03-13T07:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d to /host/opt/cni/bin/\\\\n2026-03-13T07:40:56Z [verbose] multus-daemon started\\\\n2026-03-13T07:40:56Z [verbose] Readiness Indicator file check\\\\n2026-03-13T07:41:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.794841 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.804069 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.815647 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.828130 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.843653 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.860731 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.870931 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.880177 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.893292 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.913829 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6c07029-a600-45d1-9c45-921a090a1cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a40f3daa61f0648a69b8302e64631a39a78f48a14bc0cceaf946101162c82c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d760be87350f9f6de9d09d1164320f34913b248d88551440875a177d8f2d098c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ec8a00c010220e0d0f5374f11365221798e1fea38320dd6ef538a46279632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21de48bddda9ae4ebb09f62645f9287cf815a3d342162be3a7f27209e485b85a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c022b915c81f918ec768d48e9698848b094036d5de47c50b25615faf3bfe276a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.929640 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.944406 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.960650 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.972853 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5e1f0fd-0d4f-45ed-9a9b-e988fe7a811d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa61495c58e2689c63efe08ae1a16e0665a43226d78d3b29d7dd6251b67f8194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.984291 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:50 crc kubenswrapper[4876]: I0313 07:41:50.996116 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:50Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.009874 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.058775 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6c07029-a600-45d1-9c45-921a090a1cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a40f3daa61f0648a69b8302e64631a39a78f48a14bc0cceaf946101162c82c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d760be87350f9f6de9d09d1164320f34913b248d88551440875a177d8f2d098c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ec8a00c010220e0d0f5374f11365221798e1fea38320dd6ef538a46279632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21de48bddda9ae4ebb09f62645f9287cf815a3d342162be3a7f27209e485b85a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c022b915c81f918ec768d48e9698848b094036d5de47c50b25615faf3bfe276a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.082626 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.101879 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.118311 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.135064 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5e1f0fd-0d4f-45ed-9a9b-e988fe7a811d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa61495c58e2689c63efe08ae1a16e0665a43226d78d3b29d7dd6251b67f8194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.148958 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.162959 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: E0313 07:41:51.179952 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.183777 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.197035 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.212007 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"2026-03-13T07:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d\\\\n2026-03-13T07:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d to /host/opt/cni/bin/\\\\n2026-03-13T07:40:56Z [verbose] multus-daemon started\\\\n2026-03-13T07:40:56Z [verbose] Readiness Indicator file check\\\\n2026-03-13T07:41:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.222684 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.234677 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.248725 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.263017 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.282348 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.294160 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.304718 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.315378 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.325712 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.643011 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.643077 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.643094 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.643119 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.643143 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:51Z","lastTransitionTime":"2026-03-13T07:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:51 crc kubenswrapper[4876]: E0313 07:41:51.664310 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.677289 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.677333 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.677344 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.677360 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.677372 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:51Z","lastTransitionTime":"2026-03-13T07:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:51 crc kubenswrapper[4876]: E0313 07:41:51.693526 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.698483 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.698518 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.698531 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.698548 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.698560 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:51Z","lastTransitionTime":"2026-03-13T07:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:51 crc kubenswrapper[4876]: E0313 07:41:51.718003 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.722109 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.722151 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.722166 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.722186 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.722202 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:51Z","lastTransitionTime":"2026-03-13T07:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:51 crc kubenswrapper[4876]: E0313 07:41:51.738208 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.741671 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.741721 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.741734 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.741754 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.741768 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:41:51Z","lastTransitionTime":"2026-03-13T07:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:41:51 crc kubenswrapper[4876]: E0313 07:41:51.755587 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: E0313 07:41:51.755784 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.766409 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/3.log" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.767075 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/2.log" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.773806 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" exitCode=1 Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.773854 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.773907 4876 scope.go:117] "RemoveContainer" containerID="29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.774699 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:41:51 crc kubenswrapper[4876]: E0313 07:41:51.774869 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.794335 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.804789 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.814760 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.824905 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.834696 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.844656 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"2026-03-13T07:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d\\\\n2026-03-13T07:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d to /host/opt/cni/bin/\\\\n2026-03-13T07:40:56Z [verbose] multus-daemon started\\\\n2026-03-13T07:40:56Z [verbose] Readiness Indicator file check\\\\n2026-03-13T07:41:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.854438 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.867051 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.882145 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.895307 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.908298 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.932902 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29e69c693a173e5e0d6b8adb8ccff08b87847bfe188e8d9d4ab07166758751a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:20Z\\\",\\\"message\\\":\\\"t/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0313 07:41:20.768131 7122 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-st667 in node crc\\\\nI0313 07:41:20.767969 7122 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0313 07:41:20.768139 7122 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-st667 after 0 failed attempt(s)\\\\nI0313 07:41:20.768145 7122 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-st667\\\\nF0313 07:41:20.768128 7122 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:50Z\\\",\\\"message\\\":\\\".com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821119 7452 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821325 7452 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 07:41:50.821620 7452 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821711 7452 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 07:41:50.825665 7452 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 07:41:50.825688 7452 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 07:41:50.825731 7452 ovnkube.go:599] Stopped ovnkube\\\\nI0313 07:41:50.825761 7452 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 07:41:50.825853 7452 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.949069 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.964861 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:51 crc kubenswrapper[4876]: I0313 07:41:51.977837 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5e1f0fd-0d4f-45ed-9a9b-e988fe7a811d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa61495c58e2689c63efe08ae1a16e0665a43226d78d3b29d7dd6251b67f8194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:51Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.008301 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6c07029-a600-45d1-9c45-921a090a1cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a40f3daa61f0648a69b8302e64631a39a78f48a14bc0cceaf946101162c82c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d760be87350f9f6de9d09d1164320f34913b248d88551440875a177d8f2d098c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ec8a00c010220e0d0f5374f11365221798e1fea38320dd6ef538a46279632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21de48bddda9ae4ebb09f62645f9287cf815a3d342162be3a7f27209e485b85a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c022b915c81f918ec768d48e9698848b094036d5de47c50b25615faf3bfe276a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:52Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.025517 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:52Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.035196 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.035227 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:52 crc kubenswrapper[4876]: E0313 07:41:52.035480 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.035228 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:52 crc kubenswrapper[4876]: E0313 07:41:52.035605 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:52 crc kubenswrapper[4876]: E0313 07:41:52.035694 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.035264 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:52 crc kubenswrapper[4876]: E0313 07:41:52.035990 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.045397 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:52Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.059489 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:52Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:52 crc kubenswrapper[4876]: I0313 07:41:52.779340 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/3.log" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.035296 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.035401 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.035488 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.035332 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:54 crc kubenswrapper[4876]: E0313 07:41:54.035525 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:54 crc kubenswrapper[4876]: E0313 07:41:54.035670 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:54 crc kubenswrapper[4876]: E0313 07:41:54.035758 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:54 crc kubenswrapper[4876]: E0313 07:41:54.035814 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.577829 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.578600 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:41:54 crc kubenswrapper[4876]: E0313 07:41:54.578749 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.595068 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.616399 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.631095 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"2026-03-13T07:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d\\\\n2026-03-13T07:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d to /host/opt/cni/bin/\\\\n2026-03-13T07:40:56Z [verbose] multus-daemon started\\\\n2026-03-13T07:40:56Z [verbose] Readiness Indicator file check\\\\n2026-03-13T07:41:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.643418 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.657285 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.676285 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.693128 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.713919 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:50Z\\\",\\\"message\\\":\\\".com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821119 7452 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821325 7452 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 07:41:50.821620 7452 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821711 7452 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 07:41:50.825665 7452 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 07:41:50.825688 7452 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 07:41:50.825731 7452 ovnkube.go:599] Stopped ovnkube\\\\nI0313 07:41:50.825761 7452 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 07:41:50.825853 7452 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.727894 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.740859 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.756289 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.769298 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5e1f0fd-0d4f-45ed-9a9b-e988fe7a811d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa61495c58e2689c63efe08ae1a16e0665a43226d78d3b29d7dd6251b67f8194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.793888 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6c07029-a600-45d1-9c45-921a090a1cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a40f3daa61f0648a69b8302e64631a39a78f48a14bc0cceaf946101162c82c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d760be87350f9f6de9d09d1164320f34913b248d88551440875a177d8f2d098c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ec8a00c010220e0d0f5374f11365221798e1fea38320dd6ef538a46279632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21de48bddda9ae4ebb09f62645f9287cf815a3d342162be3a7f27209e485b85a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c022b915c81f918ec768d48e9698848b094036d5de47c50b25615faf3bfe276a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.814012 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.830326 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.848338 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.870977 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.885949 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:54 crc kubenswrapper[4876]: I0313 07:41:54.898991 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:41:54Z is after 2025-08-24T17:21:41Z" Mar 13 07:41:56 crc kubenswrapper[4876]: I0313 07:41:56.035420 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:56 crc kubenswrapper[4876]: I0313 07:41:56.035514 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:56 crc kubenswrapper[4876]: E0313 07:41:56.035552 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:56 crc kubenswrapper[4876]: I0313 07:41:56.035420 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:56 crc kubenswrapper[4876]: I0313 07:41:56.035442 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:56 crc kubenswrapper[4876]: E0313 07:41:56.035642 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:56 crc kubenswrapper[4876]: E0313 07:41:56.035691 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:56 crc kubenswrapper[4876]: E0313 07:41:56.035748 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:56 crc kubenswrapper[4876]: E0313 07:41:56.180911 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.030352 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.030439 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.030466 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.030485 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.030511 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030603 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030645 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:43:02.030633338 +0000 UTC m=+241.701412320 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030694 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:43:02.03068918 +0000 UTC m=+241.701468162 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030752 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030762 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030771 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030790 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 07:43:02.030784832 +0000 UTC m=+241.701563814 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030829 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030836 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030842 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030864 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 07:43:02.030855664 +0000 UTC m=+241.701634646 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030899 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.030924 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 07:43:02.030917056 +0000 UTC m=+241.701696038 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.035481 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.035636 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.035865 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.035930 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.036062 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.036120 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.036269 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.036332 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:41:58 crc kubenswrapper[4876]: I0313 07:41:58.131267 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.131459 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:41:58 crc kubenswrapper[4876]: E0313 07:41:58.131580 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs podName:f36294fb-98b9-48be-8237-beb6484e0cb6 nodeName:}" failed. No retries permitted until 2026-03-13 07:43:02.131546238 +0000 UTC m=+241.802325260 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs") pod "network-metrics-daemon-zrbh4" (UID: "f36294fb-98b9-48be-8237-beb6484e0cb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 07:42:00 crc kubenswrapper[4876]: I0313 07:42:00.034902 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:00 crc kubenswrapper[4876]: E0313 07:42:00.035344 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:00 crc kubenswrapper[4876]: I0313 07:42:00.035080 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:00 crc kubenswrapper[4876]: E0313 07:42:00.035425 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:00 crc kubenswrapper[4876]: I0313 07:42:00.035002 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:00 crc kubenswrapper[4876]: I0313 07:42:00.035083 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:00 crc kubenswrapper[4876]: E0313 07:42:00.035483 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:00 crc kubenswrapper[4876]: E0313 07:42:00.035846 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.048743 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bac5f19-8f6a-4318-a790-d3d521079c46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c75035ec642c8dfce384b9e6ca31ede326de40044cb91149b9fa35db28e950c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ddc07348b9f4b9cff370817dd20922a04945256e4185ce249521a4a166c55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03214f4ca3878b785572d775003512dbeb92b4b36a5547b164bb57eb1bd4aaa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a3620eafcddf9c401096f361326c8992f582a305eec874238c9d1892028a2e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.061457 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.073928 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-st667" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adb2a69-d0ef-4efc-813d-77fdf9713d50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:41Z\\\",\\\"message\\\":\\\"2026-03-13T07:40:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d\\\\n2026-03-13T07:40:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a4e3654-a755-4efb-8cae-cd7f3087648d to /host/opt/cni/bin/\\\\n2026-03-13T07:40:56Z [verbose] multus-daemon started\\\\n2026-03-13T07:40:56Z [verbose] Readiness Indicator file check\\\\n2026-03-13T07:41:41Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxm4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-st667\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.087364 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f36294fb-98b9-48be-8237-beb6484e0cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjs89\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zrbh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.100278 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a6f71e5-2091-4386-b559-bba70bc45972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad08a7740efacfd72ff04e93a1bfcd9486ed1cd1953912a976015dd18e878c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmnf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r9cl2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.112045 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cec571df810c045445da8d81001796ea7f8908f4db630c91ad906cce0f847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.126218 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdedd437-30c1-4443-a0ea-1918adeb7ff4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5519774218dd89720bbe81f44770778682b3ce5bc2c672c2fd10dadceaa3b146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d9c04cd69de459fc71dda25eafa4538fe13f14bf34da3d40be0020c49a14eb2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:39:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 07:39:03.205593 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 07:39:03.208495 1 observer_polling.go:159] Starting file observer\\\\nI0313 07:39:03.237257 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 07:39:03.240272 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 07:39:32.036273 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 07:39:32.036357 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:39:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0afe3c4e823c8d0d5d3ff5b4e847c0af3599ef1fa3b7209368472318beff30f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c117e498de2d7f8fee01a0163bcc8d09fe330ea5dd451a06a0290a0d51415d03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.144349 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2acda0f0-90e3-4168-9754-abd0bc4bf37b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T07:40:11Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0313 07:40:11.786294 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0313 07:40:11.786335 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI0313 07:40:11.786427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0313 07:40:11.786448 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0313 07:40:11.786520 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\"\\\\nI0313 07:40:11.786536 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1016347625/tls.crt::/tmp/serving-cert-1016347625/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1773387611\\\\\\\\\\\\\\\" (2026-03-13 07:40:10 +0000 UTC to 2026-04-12 07:40:11 +0000 UTC (now=2026-03-13 07:40:11.786504427 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786679 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1773387611\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1773387611\\\\\\\\\\\\\\\" (2026-03-13 06:40:11 +0000 UTC to 2027-03-13 06:40:11 +0000 UTC (now=2026-03-13 07:40:11.786657732 +0000 UTC))\\\\\\\"\\\\nI0313 07:40:11.786714 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0313 07:40:11.786735 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0313 07:40:11.786793 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0313 07:40:11.787572 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.159200 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdbbbfb70542a8657088575999bb4a24e16ce4dd81474fdbbf0f28ec2d8c4585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: E0313 07:42:01.181405 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.182679 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T07:41:50Z\\\",\\\"message\\\":\\\".com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821119 7452 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821325 7452 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 07:41:50.821620 7452 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 07:41:50.821711 7452 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 07:41:50.825665 7452 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 07:41:50.825688 7452 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 07:41:50.825731 7452 ovnkube.go:599] Stopped ovnkube\\\\nI0313 07:41:50.825761 7452 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 07:41:50.825853 7452 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T07:41:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-g4xkn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.199319 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.211324 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5e1f0fd-0d4f-45ed-9a9b-e988fe7a811d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa61495c58e2689c63efe08ae1a16e0665a43226d78d3b29d7dd6251b67f8194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4eeef8661302363dc915bae1312372fce759616a03af57a059b6faf6bacf64fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.228867 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6c07029-a600-45d1-9c45-921a090a1cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:39:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a40f3daa61f0648a69b8302e64631a39a78f48a14bc0cceaf946101162c82c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d760be87350f9f6de9d09d1164320f34913b248d88551440875a177d8f2d098c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ec8a00c010220e0d0f5374f11365221798e1fea38320dd6ef538a46279632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21de48bddda9ae4ebb09f62645f9287cf815a3d342162be3a7f27209e485b85a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c022b915c81f918ec768d48e9698848b094036d5de47c50b25615faf3bfe276a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:39:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecf3c970748f5f2721e9c83b0500079c94fc169b24d3cf781646143e9e1e025f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc7960fcc9cd42bf6f2ceef5aaef7e7fb0c36d0103c54af2060100771e371e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:03Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0fc79d0d6355f33892937c618f22bd9de5e1905fdcf589cac23e81396603d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:39:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:39:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:39:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.243553 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gqsld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"153920eb-fdb1-4e94-85ca-f09b5f01c2aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbad2b918c1f90d27dad2bad86452469d28f7b13e5f3e6ddfdf9106573c63f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b836d6bfafd7155a751b82d3e849a10d241a86780f949ec80efd2d6fd1da3251\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a921c9b9af2024a0e2e93847c01fdecf554f78c0293a643eb66f1ba138d023a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4598e245ab6daba35ae4adaa1f488763afead5c8af57409e67097875c0b5fa3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2be316de8512eb6556f287ff212d0a766b2f0ea920fef1bcadd04e1d19eb3a73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d909f266d6ae764e5c6e6d89d0ef0f17a1372ccbce2d312ebc0484248bc32bf0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5eb394e6c952eeff55f065d044e9f7af7a8ed3d2e45166f0359901972503c04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T07:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T07:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-smtrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gqsld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.255685 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc05487d639440dd390716799b02300aff8145a614cd4fee82035fbd2a27a217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a91d8843c4987a517ac7c50394157db0b0893d0e36d7946bb24d6eea36fe7a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.266613 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdd10539-ff2d-4055-86eb-d422e532bf66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20f06fb7fc41532b467951aa5c1e074a45b25e07b375bd20e714d947ef1ffc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ac5a79efb96a16b2caf3e45e74d5a19bd4558837ab9fff9cff4bdc230d9acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgt4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-59l8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.278116 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.287458 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-sz689" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"182bf6da-086a-4b4e-b69c-411b5f6e8593\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://048e0608682d059da7f40d52060115f02feb407695c5f321385163f1b0745e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxrm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-sz689\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.297805 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-krrjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b230d31c-acdd-44a2-85b5-30e786f515c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T07:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7bfd5f58d2f7039c0f1b41ca4a5c954502863e9a66baca3d713819084cdafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T07:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-twfww\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T07:40:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-krrjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.850673 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.850713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.850722 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.850760 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.850771 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:42:01Z","lastTransitionTime":"2026-03-13T07:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:42:01 crc kubenswrapper[4876]: E0313 07:42:01.869529 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.873557 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.873630 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.873656 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.873689 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.873715 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:42:01Z","lastTransitionTime":"2026-03-13T07:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:42:01 crc kubenswrapper[4876]: E0313 07:42:01.887211 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.890853 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.890887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.890896 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.890910 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.890920 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:42:01Z","lastTransitionTime":"2026-03-13T07:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:42:01 crc kubenswrapper[4876]: E0313 07:42:01.906827 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.909945 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.909979 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.909990 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.910004 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.910014 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:42:01Z","lastTransitionTime":"2026-03-13T07:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:42:01 crc kubenswrapper[4876]: E0313 07:42:01.920732 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.924210 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.924270 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.924281 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.924297 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:42:01 crc kubenswrapper[4876]: I0313 07:42:01.924308 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:42:01Z","lastTransitionTime":"2026-03-13T07:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:42:01 crc kubenswrapper[4876]: E0313 07:42:01.936164 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T07:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b24a1e5a-8f3e-43aa-b425-2a9854971c62\\\",\\\"systemUUID\\\":\\\"b1858cc2-71ca-4a24-ba44-14334b24f2dd\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T07:42:01Z is after 2025-08-24T17:21:41Z" Mar 13 07:42:01 crc kubenswrapper[4876]: E0313 07:42:01.936334 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 07:42:02 crc kubenswrapper[4876]: I0313 07:42:02.035285 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:02 crc kubenswrapper[4876]: I0313 07:42:02.035285 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:02 crc kubenswrapper[4876]: I0313 07:42:02.035384 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:02 crc kubenswrapper[4876]: I0313 07:42:02.035418 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:02 crc kubenswrapper[4876]: E0313 07:42:02.035600 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:02 crc kubenswrapper[4876]: E0313 07:42:02.035729 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:02 crc kubenswrapper[4876]: E0313 07:42:02.035827 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:02 crc kubenswrapper[4876]: E0313 07:42:02.035890 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:04 crc kubenswrapper[4876]: I0313 07:42:04.034900 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:04 crc kubenswrapper[4876]: I0313 07:42:04.034969 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:04 crc kubenswrapper[4876]: E0313 07:42:04.035026 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:04 crc kubenswrapper[4876]: E0313 07:42:04.035116 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:04 crc kubenswrapper[4876]: I0313 07:42:04.035132 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:04 crc kubenswrapper[4876]: E0313 07:42:04.035232 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:04 crc kubenswrapper[4876]: I0313 07:42:04.035305 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:04 crc kubenswrapper[4876]: E0313 07:42:04.035559 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:06 crc kubenswrapper[4876]: I0313 07:42:06.035061 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:06 crc kubenswrapper[4876]: I0313 07:42:06.035125 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:06 crc kubenswrapper[4876]: I0313 07:42:06.035127 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:06 crc kubenswrapper[4876]: I0313 07:42:06.035150 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:06 crc kubenswrapper[4876]: E0313 07:42:06.035429 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:06 crc kubenswrapper[4876]: E0313 07:42:06.036127 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:06 crc kubenswrapper[4876]: E0313 07:42:06.036453 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:06 crc kubenswrapper[4876]: E0313 07:42:06.036559 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:06 crc kubenswrapper[4876]: I0313 07:42:06.036984 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:42:06 crc kubenswrapper[4876]: E0313 07:42:06.037365 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:42:06 crc kubenswrapper[4876]: E0313 07:42:06.183365 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:42:08 crc kubenswrapper[4876]: I0313 07:42:08.034975 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:08 crc kubenswrapper[4876]: E0313 07:42:08.035109 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:08 crc kubenswrapper[4876]: I0313 07:42:08.035165 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:08 crc kubenswrapper[4876]: I0313 07:42:08.035176 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:08 crc kubenswrapper[4876]: E0313 07:42:08.035568 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:08 crc kubenswrapper[4876]: E0313 07:42:08.035618 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:08 crc kubenswrapper[4876]: I0313 07:42:08.035853 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:08 crc kubenswrapper[4876]: E0313 07:42:08.035964 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:10 crc kubenswrapper[4876]: I0313 07:42:10.034838 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:10 crc kubenswrapper[4876]: I0313 07:42:10.034895 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:10 crc kubenswrapper[4876]: I0313 07:42:10.034890 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:10 crc kubenswrapper[4876]: E0313 07:42:10.035971 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:10 crc kubenswrapper[4876]: E0313 07:42:10.035725 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:10 crc kubenswrapper[4876]: I0313 07:42:10.035270 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:10 crc kubenswrapper[4876]: E0313 07:42:10.036168 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:10 crc kubenswrapper[4876]: E0313 07:42:10.036193 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.074494 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-st667" podStartSLOduration=132.074475527 podStartE2EDuration="2m12.074475527s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.074369914 +0000 UTC m=+190.745148946" watchObservedRunningTime="2026-03-13 07:42:11.074475527 +0000 UTC m=+190.745254509" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.119444 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=68.119430146 podStartE2EDuration="1m8.119430146s" podCreationTimestamp="2026-03-13 07:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.10399778 +0000 UTC m=+190.774776762" watchObservedRunningTime="2026-03-13 07:42:11.119430146 +0000 UTC m=+190.790209118" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.132429 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.132411472 podStartE2EDuration="1m17.132411472s" podCreationTimestamp="2026-03-13 07:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.120712342 +0000 UTC m=+190.791491324" watchObservedRunningTime="2026-03-13 07:42:11.132411472 +0000 UTC m=+190.803190454" Mar 13 07:42:11 crc kubenswrapper[4876]: E0313 07:42:11.183865 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.207040 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podStartSLOduration=132.20702222 podStartE2EDuration="2m12.20702222s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.189494494 +0000 UTC m=+190.860273476" watchObservedRunningTime="2026-03-13 07:42:11.20702222 +0000 UTC m=+190.877801202" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.227783 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=63.227763366 podStartE2EDuration="1m3.227763366s" podCreationTimestamp="2026-03-13 07:41:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.227271811 +0000 UTC m=+190.898050803" watchObservedRunningTime="2026-03-13 07:42:11.227763366 +0000 UTC m=+190.898542348" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.259162 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=27.259146291 podStartE2EDuration="27.259146291s" podCreationTimestamp="2026-03-13 07:41:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.258059701 +0000 UTC m=+190.928838683" watchObservedRunningTime="2026-03-13 07:42:11.259146291 +0000 UTC m=+190.929925273" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.274172 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-gqsld" podStartSLOduration=132.274156466 podStartE2EDuration="2m12.274156466s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.27361181 +0000 UTC m=+190.944390792" watchObservedRunningTime="2026-03-13 07:42:11.274156466 +0000 UTC m=+190.944935448" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.296553 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-59l8k" podStartSLOduration=131.296535298 podStartE2EDuration="2m11.296535298s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.296183497 +0000 UTC m=+190.966962499" watchObservedRunningTime="2026-03-13 07:42:11.296535298 +0000 UTC m=+190.967314280" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.304740 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.304724089 podStartE2EDuration="21.304724089s" podCreationTimestamp="2026-03-13 07:41:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.304705488 +0000 UTC m=+190.975484470" watchObservedRunningTime="2026-03-13 07:42:11.304724089 +0000 UTC m=+190.975503061" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.314347 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-sz689" podStartSLOduration=132.314329639 podStartE2EDuration="2m12.314329639s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.313806085 +0000 UTC m=+190.984585057" watchObservedRunningTime="2026-03-13 07:42:11.314329639 +0000 UTC m=+190.985108621" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.336541 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-krrjs" podStartSLOduration=132.336523406 podStartE2EDuration="2m12.336523406s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:11.324698253 +0000 UTC m=+190.995477235" watchObservedRunningTime="2026-03-13 07:42:11.336523406 +0000 UTC m=+191.007302388" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.959144 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.959201 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.959212 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.959227 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 07:42:11 crc kubenswrapper[4876]: I0313 07:42:11.959261 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T07:42:11Z","lastTransitionTime":"2026-03-13T07:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.016892 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx"] Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.017541 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.019831 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.022842 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.023129 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.023168 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.034571 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:12 crc kubenswrapper[4876]: E0313 07:42:12.034742 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.034999 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:12 crc kubenswrapper[4876]: E0313 07:42:12.035089 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.035265 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.035630 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:12 crc kubenswrapper[4876]: E0313 07:42:12.035708 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:12 crc kubenswrapper[4876]: E0313 07:42:12.036182 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.176866 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cc6c203-870c-4258-8459-37b28c94a924-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.176949 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cc6c203-870c-4258-8459-37b28c94a924-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.177040 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cc6c203-870c-4258-8459-37b28c94a924-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.177093 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cc6c203-870c-4258-8459-37b28c94a924-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.177127 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cc6c203-870c-4258-8459-37b28c94a924-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.220898 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.230744 4876 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.278355 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cc6c203-870c-4258-8459-37b28c94a924-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.278452 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cc6c203-870c-4258-8459-37b28c94a924-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.278494 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cc6c203-870c-4258-8459-37b28c94a924-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.278552 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cc6c203-870c-4258-8459-37b28c94a924-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.278641 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cc6c203-870c-4258-8459-37b28c94a924-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.278795 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cc6c203-870c-4258-8459-37b28c94a924-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.279013 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cc6c203-870c-4258-8459-37b28c94a924-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.280500 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cc6c203-870c-4258-8459-37b28c94a924-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.290106 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cc6c203-870c-4258-8459-37b28c94a924-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.310722 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cc6c203-870c-4258-8459-37b28c94a924-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4w4jx\" (UID: \"0cc6c203-870c-4258-8459-37b28c94a924\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.346121 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.848481 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" event={"ID":"0cc6c203-870c-4258-8459-37b28c94a924","Type":"ContainerStarted","Data":"fbcc17f4133dc16f2c997e37acc587d42ef387178614cacaad88229b4471c415"} Mar 13 07:42:12 crc kubenswrapper[4876]: I0313 07:42:12.848894 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" event={"ID":"0cc6c203-870c-4258-8459-37b28c94a924","Type":"ContainerStarted","Data":"3f302711e33b0bb58164e8e2f32b266f6c276330b9cf4e919130433eacb86aec"} Mar 13 07:42:14 crc kubenswrapper[4876]: I0313 07:42:14.034870 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:14 crc kubenswrapper[4876]: I0313 07:42:14.034962 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:14 crc kubenswrapper[4876]: I0313 07:42:14.035035 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:14 crc kubenswrapper[4876]: E0313 07:42:14.035035 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:14 crc kubenswrapper[4876]: I0313 07:42:14.035094 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:14 crc kubenswrapper[4876]: E0313 07:42:14.035299 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:14 crc kubenswrapper[4876]: E0313 07:42:14.035404 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:14 crc kubenswrapper[4876]: E0313 07:42:14.035509 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:16 crc kubenswrapper[4876]: I0313 07:42:16.034840 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:16 crc kubenswrapper[4876]: I0313 07:42:16.034865 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:16 crc kubenswrapper[4876]: I0313 07:42:16.034893 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:16 crc kubenswrapper[4876]: E0313 07:42:16.034985 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:16 crc kubenswrapper[4876]: E0313 07:42:16.035149 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:16 crc kubenswrapper[4876]: E0313 07:42:16.035255 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:16 crc kubenswrapper[4876]: I0313 07:42:16.035326 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:16 crc kubenswrapper[4876]: E0313 07:42:16.035392 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:16 crc kubenswrapper[4876]: E0313 07:42:16.184796 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:42:18 crc kubenswrapper[4876]: I0313 07:42:18.035559 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:18 crc kubenswrapper[4876]: I0313 07:42:18.035668 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:18 crc kubenswrapper[4876]: I0313 07:42:18.035622 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:18 crc kubenswrapper[4876]: E0313 07:42:18.035844 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:18 crc kubenswrapper[4876]: E0313 07:42:18.036146 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:18 crc kubenswrapper[4876]: E0313 07:42:18.036317 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:18 crc kubenswrapper[4876]: I0313 07:42:18.035966 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:18 crc kubenswrapper[4876]: E0313 07:42:18.036761 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:20 crc kubenswrapper[4876]: I0313 07:42:20.035366 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:20 crc kubenswrapper[4876]: I0313 07:42:20.035437 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:20 crc kubenswrapper[4876]: I0313 07:42:20.035477 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:20 crc kubenswrapper[4876]: E0313 07:42:20.035506 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:20 crc kubenswrapper[4876]: I0313 07:42:20.035597 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:20 crc kubenswrapper[4876]: E0313 07:42:20.035658 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:20 crc kubenswrapper[4876]: E0313 07:42:20.035743 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:20 crc kubenswrapper[4876]: E0313 07:42:20.035790 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:20 crc kubenswrapper[4876]: I0313 07:42:20.036400 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:42:20 crc kubenswrapper[4876]: E0313 07:42:20.036633 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-g4xkn_openshift-ovn-kubernetes(6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" Mar 13 07:42:21 crc kubenswrapper[4876]: E0313 07:42:21.185439 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:42:22 crc kubenswrapper[4876]: I0313 07:42:22.035076 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:22 crc kubenswrapper[4876]: I0313 07:42:22.035132 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:22 crc kubenswrapper[4876]: I0313 07:42:22.035188 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:22 crc kubenswrapper[4876]: E0313 07:42:22.035342 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:22 crc kubenswrapper[4876]: I0313 07:42:22.035446 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:22 crc kubenswrapper[4876]: E0313 07:42:22.035537 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:22 crc kubenswrapper[4876]: E0313 07:42:22.035691 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:22 crc kubenswrapper[4876]: E0313 07:42:22.035938 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:24 crc kubenswrapper[4876]: I0313 07:42:24.034625 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:24 crc kubenswrapper[4876]: I0313 07:42:24.034659 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:24 crc kubenswrapper[4876]: I0313 07:42:24.034668 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:24 crc kubenswrapper[4876]: E0313 07:42:24.034775 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:24 crc kubenswrapper[4876]: I0313 07:42:24.034803 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:24 crc kubenswrapper[4876]: E0313 07:42:24.034909 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:24 crc kubenswrapper[4876]: E0313 07:42:24.035028 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:24 crc kubenswrapper[4876]: E0313 07:42:24.034966 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:26 crc kubenswrapper[4876]: I0313 07:42:26.034787 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:26 crc kubenswrapper[4876]: I0313 07:42:26.034853 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:26 crc kubenswrapper[4876]: E0313 07:42:26.034926 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:26 crc kubenswrapper[4876]: I0313 07:42:26.035014 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:26 crc kubenswrapper[4876]: E0313 07:42:26.035054 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:26 crc kubenswrapper[4876]: I0313 07:42:26.035081 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:26 crc kubenswrapper[4876]: E0313 07:42:26.039185 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:26 crc kubenswrapper[4876]: E0313 07:42:26.039426 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:26 crc kubenswrapper[4876]: E0313 07:42:26.186924 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:42:27 crc kubenswrapper[4876]: I0313 07:42:27.896380 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/1.log" Mar 13 07:42:27 crc kubenswrapper[4876]: I0313 07:42:27.898068 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/0.log" Mar 13 07:42:27 crc kubenswrapper[4876]: I0313 07:42:27.898153 4876 generic.go:334] "Generic (PLEG): container finished" podID="8adb2a69-d0ef-4efc-813d-77fdf9713d50" containerID="a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4" exitCode=1 Mar 13 07:42:27 crc kubenswrapper[4876]: I0313 07:42:27.898222 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-st667" event={"ID":"8adb2a69-d0ef-4efc-813d-77fdf9713d50","Type":"ContainerDied","Data":"a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4"} Mar 13 07:42:27 crc kubenswrapper[4876]: I0313 07:42:27.898357 4876 scope.go:117] "RemoveContainer" containerID="77fba5651a4d48fa091f5bfffa7e61988b3af260308ba43626914cc1c7335369" Mar 13 07:42:27 crc kubenswrapper[4876]: I0313 07:42:27.898978 4876 scope.go:117] "RemoveContainer" containerID="a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4" Mar 13 07:42:27 crc kubenswrapper[4876]: E0313 07:42:27.899317 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-st667_openshift-multus(8adb2a69-d0ef-4efc-813d-77fdf9713d50)\"" pod="openshift-multus/multus-st667" podUID="8adb2a69-d0ef-4efc-813d-77fdf9713d50" Mar 13 07:42:27 crc kubenswrapper[4876]: I0313 07:42:27.927353 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4w4jx" podStartSLOduration=148.927333839 podStartE2EDuration="2m28.927333839s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:12.8676017 +0000 UTC m=+192.538380712" watchObservedRunningTime="2026-03-13 07:42:27.927333839 +0000 UTC m=+207.598112821" Mar 13 07:42:28 crc kubenswrapper[4876]: I0313 07:42:28.034437 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:28 crc kubenswrapper[4876]: E0313 07:42:28.034557 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:28 crc kubenswrapper[4876]: I0313 07:42:28.034605 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:28 crc kubenswrapper[4876]: I0313 07:42:28.034671 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:28 crc kubenswrapper[4876]: E0313 07:42:28.034725 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:28 crc kubenswrapper[4876]: E0313 07:42:28.034764 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:28 crc kubenswrapper[4876]: I0313 07:42:28.034915 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:28 crc kubenswrapper[4876]: E0313 07:42:28.035094 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:28 crc kubenswrapper[4876]: I0313 07:42:28.903894 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/1.log" Mar 13 07:42:30 crc kubenswrapper[4876]: I0313 07:42:30.035292 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:30 crc kubenswrapper[4876]: I0313 07:42:30.035337 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:30 crc kubenswrapper[4876]: E0313 07:42:30.035401 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:30 crc kubenswrapper[4876]: I0313 07:42:30.035307 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:30 crc kubenswrapper[4876]: E0313 07:42:30.035526 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:30 crc kubenswrapper[4876]: E0313 07:42:30.035595 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:30 crc kubenswrapper[4876]: I0313 07:42:30.036334 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:30 crc kubenswrapper[4876]: E0313 07:42:30.036512 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:31 crc kubenswrapper[4876]: E0313 07:42:31.187624 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.035385 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.035422 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.035471 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:32 crc kubenswrapper[4876]: E0313 07:42:32.035584 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.035596 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.036387 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:42:32 crc kubenswrapper[4876]: E0313 07:42:32.036576 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:32 crc kubenswrapper[4876]: E0313 07:42:32.036652 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:32 crc kubenswrapper[4876]: E0313 07:42:32.036587 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.919004 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/3.log" Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.921857 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerStarted","Data":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.922403 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:42:32 crc kubenswrapper[4876]: I0313 07:42:32.965092 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podStartSLOduration=152.965052872 podStartE2EDuration="2m32.965052872s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:32.964554037 +0000 UTC m=+212.635333039" watchObservedRunningTime="2026-03-13 07:42:32.965052872 +0000 UTC m=+212.635831854" Mar 13 07:42:33 crc kubenswrapper[4876]: I0313 07:42:33.228266 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zrbh4"] Mar 13 07:42:33 crc kubenswrapper[4876]: I0313 07:42:33.228371 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:33 crc kubenswrapper[4876]: E0313 07:42:33.228520 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:34 crc kubenswrapper[4876]: I0313 07:42:34.034797 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:34 crc kubenswrapper[4876]: I0313 07:42:34.034910 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:34 crc kubenswrapper[4876]: I0313 07:42:34.034924 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:34 crc kubenswrapper[4876]: E0313 07:42:34.035314 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:34 crc kubenswrapper[4876]: E0313 07:42:34.035374 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:34 crc kubenswrapper[4876]: E0313 07:42:34.035396 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:35 crc kubenswrapper[4876]: I0313 07:42:35.035218 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:35 crc kubenswrapper[4876]: E0313 07:42:35.035409 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:36 crc kubenswrapper[4876]: I0313 07:42:36.035227 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:36 crc kubenswrapper[4876]: I0313 07:42:36.035357 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:36 crc kubenswrapper[4876]: E0313 07:42:36.035474 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:36 crc kubenswrapper[4876]: I0313 07:42:36.035521 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:36 crc kubenswrapper[4876]: E0313 07:42:36.035600 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:36 crc kubenswrapper[4876]: E0313 07:42:36.035648 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:36 crc kubenswrapper[4876]: E0313 07:42:36.189108 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 07:42:37 crc kubenswrapper[4876]: I0313 07:42:37.035462 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:37 crc kubenswrapper[4876]: E0313 07:42:37.035626 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:38 crc kubenswrapper[4876]: I0313 07:42:38.034838 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:38 crc kubenswrapper[4876]: I0313 07:42:38.034878 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:38 crc kubenswrapper[4876]: I0313 07:42:38.034878 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:38 crc kubenswrapper[4876]: E0313 07:42:38.035010 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:38 crc kubenswrapper[4876]: E0313 07:42:38.035094 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:38 crc kubenswrapper[4876]: E0313 07:42:38.035214 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:39 crc kubenswrapper[4876]: I0313 07:42:39.035143 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:39 crc kubenswrapper[4876]: E0313 07:42:39.035408 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:39 crc kubenswrapper[4876]: I0313 07:42:39.035756 4876 scope.go:117] "RemoveContainer" containerID="a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4" Mar 13 07:42:39 crc kubenswrapper[4876]: I0313 07:42:39.944263 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/1.log" Mar 13 07:42:39 crc kubenswrapper[4876]: I0313 07:42:39.944322 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-st667" event={"ID":"8adb2a69-d0ef-4efc-813d-77fdf9713d50","Type":"ContainerStarted","Data":"76512308b59931886eee75e5bd8c0ff26eed1bfe288f2ff41e80a8ca4761ad73"} Mar 13 07:42:40 crc kubenswrapper[4876]: I0313 07:42:40.035218 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:40 crc kubenswrapper[4876]: I0313 07:42:40.035260 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:40 crc kubenswrapper[4876]: I0313 07:42:40.035323 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:40 crc kubenswrapper[4876]: E0313 07:42:40.035339 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 07:42:40 crc kubenswrapper[4876]: E0313 07:42:40.036716 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 07:42:40 crc kubenswrapper[4876]: E0313 07:42:40.036800 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 07:42:41 crc kubenswrapper[4876]: I0313 07:42:41.035426 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:41 crc kubenswrapper[4876]: E0313 07:42:41.037389 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zrbh4" podUID="f36294fb-98b9-48be-8237-beb6484e0cb6" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.035408 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.035517 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.035644 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.038639 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.038675 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.038643 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.039054 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.896587 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.938373 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nkhkk"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.938765 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.941066 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xvwj2"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.941599 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.943979 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.944143 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.950558 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.951405 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.951623 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.951750 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.955027 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.955041 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.958312 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.960124 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.960357 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.960468 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.960615 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.961598 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.962826 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.964978 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.965517 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.965794 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.965877 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.967314 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.976340 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6qqdn"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.976936 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.977019 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.977152 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.977271 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xvfbt"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.977404 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.977432 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.977534 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.978105 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.978266 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-g65sf"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.978709 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.980783 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lqr7r"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.981339 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.981708 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nhkg4"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.982430 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.983461 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.983889 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.988238 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.988551 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.988793 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.988852 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.988970 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989001 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.988803 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989094 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989396 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989436 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989556 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989651 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989715 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989735 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989810 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.989908 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.990011 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.990055 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.990119 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5rhkc"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.990479 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.990537 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-cd4ct"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.990699 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.990818 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.991046 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.991164 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.991299 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.991313 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.991414 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.992117 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr"] Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.992328 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.992456 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.992848 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.992890 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.992992 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993083 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993126 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993155 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993222 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993335 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993413 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993483 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993535 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993651 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993790 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993821 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993891 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993901 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993981 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993995 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.994026 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993798 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.994071 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993484 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.994125 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.994143 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.993999 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.994224 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.994326 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.994464 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.994566 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.995370 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.995477 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.995497 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.995596 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.995700 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.995817 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.996065 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.996131 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.996212 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.996273 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 07:42:42 crc kubenswrapper[4876]: I0313 07:42:42.996075 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.000621 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.016434 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.017676 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.027589 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.027587 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6cnvw"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.031483 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.037604 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.040065 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.040211 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.054339 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.056125 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.059480 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.062465 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.066721 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.066962 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.067416 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tpr6j"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.067910 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.068205 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.068738 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.068762 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.068878 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.069317 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.069808 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5pvl"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.070178 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.070551 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.070575 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.070712 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.070767 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.071280 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.071424 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.072718 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.073079 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.083079 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.083680 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.083995 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.084200 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.084352 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.085099 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.088258 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lw2ss"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.088910 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-75kbq"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.089345 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.089758 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.090191 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.091051 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.091376 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.091991 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.092283 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.093413 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094222 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094277 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5461c713-e58f-46e8-91f7-5d6546011e27-node-pullsecrets\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094296 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-serving-cert\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094313 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/945666f1-2703-4fa7-9063-fd49b7189341-config\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094329 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/418bec94-3753-4d76-b2a8-24b04f39c820-images\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094346 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-config\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094363 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-client\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094378 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094395 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g2b2\" (UniqueName: \"kubernetes.io/projected/418bec94-3753-4d76-b2a8-24b04f39c820-kube-api-access-6g2b2\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094410 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094424 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094438 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-serving-cert\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094452 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094469 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094484 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-encryption-config\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094500 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-config\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094513 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-serving-cert\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094529 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h49q\" (UniqueName: \"kubernetes.io/projected/918fce5c-2c8a-46a8-9c16-3911fb36faa3-kube-api-access-6h49q\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094545 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99f7833f-00a4-45fe-80ba-6c22404c49e4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094558 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-etcd-client\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094573 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-client-ca\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094586 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-client-ca\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094602 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-audit\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094616 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094631 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99f7833f-00a4-45fe-80ba-6c22404c49e4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094645 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99f7833f-00a4-45fe-80ba-6c22404c49e4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094668 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-config\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094683 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094697 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094710 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418bec94-3753-4d76-b2a8-24b04f39c820-config\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094726 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-trusted-ca-bundle\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094740 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/945666f1-2703-4fa7-9063-fd49b7189341-trusted-ca\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094754 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-dir\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094770 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdzcp\" (UniqueName: \"kubernetes.io/projected/945666f1-2703-4fa7-9063-fd49b7189341-kube-api-access-gdzcp\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094786 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-config\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094802 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094816 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1fe6fb32-9851-48de-88f6-94701fcf7010-service-ca-bundle\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094832 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n6t6\" (UniqueName: \"kubernetes.io/projected/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-kube-api-access-4n6t6\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094848 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-serving-cert\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094862 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caab36f1-979c-459d-b742-941e13dc765e-serving-cert\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094876 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918fce5c-2c8a-46a8-9c16-3911fb36faa3-serving-cert\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094890 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-ca\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094904 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-etcd-serving-ca\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094920 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a8ea6bd-3402-40ee-b509-51c54f8af982-serving-cert\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094935 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-service-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094955 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-serving-cert\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094962 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094968 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c59f6\" (UniqueName: \"kubernetes.io/projected/62fdbab2-2be8-42cc-a936-a87ffd085a0e-kube-api-access-c59f6\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.094990 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-service-ca\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095004 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095031 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr74m\" (UniqueName: \"kubernetes.io/projected/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-kube-api-access-sr74m\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095048 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095064 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-etcd-client\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095077 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095093 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72bdr\" (UniqueName: \"kubernetes.io/projected/99f7833f-00a4-45fe-80ba-6c22404c49e4-kube-api-access-72bdr\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095108 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-metrics-certs\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095128 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-config\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095141 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095155 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-config\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095169 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-policies\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095192 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-audit-policies\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095205 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-oauth-serving-cert\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095219 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkcd2\" (UniqueName: \"kubernetes.io/projected/ee3187e1-6603-4cd1-a134-6437fae1ca4a-kube-api-access-fkcd2\") pod \"downloads-7954f5f757-6cnvw\" (UID: \"ee3187e1-6603-4cd1-a134-6437fae1ca4a\") " pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095252 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095269 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cttzt\" (UniqueName: \"kubernetes.io/projected/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-kube-api-access-cttzt\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095283 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-image-import-ca\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095298 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97r5d\" (UniqueName: \"kubernetes.io/projected/caab36f1-979c-459d-b742-941e13dc765e-kube-api-access-97r5d\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095315 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095329 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5461c713-e58f-46e8-91f7-5d6546011e27-audit-dir\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095345 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/418bec94-3753-4d76-b2a8-24b04f39c820-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095368 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-service-ca\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095382 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/733daf88-a275-4be5-ace0-0d12edaf1b87-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095400 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf48g\" (UniqueName: \"kubernetes.io/projected/6b92045d-38c7-4dff-be0a-78449d870bcb-kube-api-access-mf48g\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095414 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drh79\" (UniqueName: \"kubernetes.io/projected/5e0a197c-3a94-4426-a492-d541a18194a0-kube-api-access-drh79\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095430 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-config\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095444 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szj2c\" (UniqueName: \"kubernetes.io/projected/1fe6fb32-9851-48de-88f6-94701fcf7010-kube-api-access-szj2c\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095459 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095473 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/945666f1-2703-4fa7-9063-fd49b7189341-serving-cert\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095487 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/733daf88-a275-4be5-ace0-0d12edaf1b87-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095503 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62jbp\" (UniqueName: \"kubernetes.io/projected/0a8ea6bd-3402-40ee-b509-51c54f8af982-kube-api-access-62jbp\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095518 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-stats-auth\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095532 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095621 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e0a197c-3a94-4426-a492-d541a18194a0-audit-dir\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095638 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095653 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-encryption-config\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095668 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-oauth-config\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095682 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/caab36f1-979c-459d-b742-941e13dc765e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095698 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-default-certificate\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095715 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.095729 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/733daf88-a275-4be5-ace0-0d12edaf1b87-config\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.096139 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.109054 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.096989 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.097250 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.112169 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.112952 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt6bn\" (UniqueName: \"kubernetes.io/projected/5461c713-e58f-46e8-91f7-5d6546011e27-kube-api-access-dt6bn\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.113408 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.116911 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.120736 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.121905 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556462-xxthz"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.122745 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556462-xxthz" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.126903 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.135176 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.136463 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.144418 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.145476 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.145790 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.146831 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6k6wd"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.147008 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.147446 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.147789 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.148196 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.149272 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.149967 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.150307 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.152588 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nkhkk"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.160326 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2bszj"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.160940 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xvwj2"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.161016 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.161762 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.164016 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.168932 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5rhkc"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.170604 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.172608 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.175127 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nhkg4"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.178203 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.179349 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6cnvw"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.180429 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.180550 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xvfbt"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.182147 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lqr7r"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.184204 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dfkkx"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.185127 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.185222 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.185402 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.188228 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.189302 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.190320 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5pvl"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.192090 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.193441 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m2gxf"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.194718 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.194955 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.197175 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556462-xxthz"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.198489 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.199988 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.200843 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.201326 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.202351 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lw2ss"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.205501 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.206872 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.208074 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-g65sf"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.210355 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.211778 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6qqdn"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214312 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-metrics-certs\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214355 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-config\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214382 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-policies\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214428 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd2j2\" (UniqueName: \"kubernetes.io/projected/300d0a76-b532-4124-af3d-f89cf7a51d3f-kube-api-access-vd2j2\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214464 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-config\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214484 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214507 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f60bac9-dca3-4eb8-b163-8146b084e74d-proxy-tls\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214531 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkcd2\" (UniqueName: \"kubernetes.io/projected/ee3187e1-6603-4cd1-a134-6437fae1ca4a-kube-api-access-fkcd2\") pod \"downloads-7954f5f757-6cnvw\" (UID: \"ee3187e1-6603-4cd1-a134-6437fae1ca4a\") " pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214553 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-audit-policies\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214579 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-oauth-serving-cert\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214602 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-image-import-ca\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214626 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97r5d\" (UniqueName: \"kubernetes.io/projected/caab36f1-979c-459d-b742-941e13dc765e-kube-api-access-97r5d\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214648 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2037c5ee-5855-4520-af01-000ac44338bb-signing-cabundle\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214673 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214699 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cttzt\" (UniqueName: \"kubernetes.io/projected/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-kube-api-access-cttzt\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214724 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/418bec94-3753-4d76-b2a8-24b04f39c820-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214747 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-service-ca\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214769 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/733daf88-a275-4be5-ace0-0d12edaf1b87-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214796 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214817 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5461c713-e58f-46e8-91f7-5d6546011e27-audit-dir\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214840 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf48g\" (UniqueName: \"kubernetes.io/projected/6b92045d-38c7-4dff-be0a-78449d870bcb-kube-api-access-mf48g\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214865 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drh79\" (UniqueName: \"kubernetes.io/projected/5e0a197c-3a94-4426-a492-d541a18194a0-kube-api-access-drh79\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214889 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-config\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214910 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szj2c\" (UniqueName: \"kubernetes.io/projected/1fe6fb32-9851-48de-88f6-94701fcf7010-kube-api-access-szj2c\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214932 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/733daf88-a275-4be5-ace0-0d12edaf1b87-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214958 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2037c5ee-5855-4520-af01-000ac44338bb-signing-key\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.214981 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62jbp\" (UniqueName: \"kubernetes.io/projected/0a8ea6bd-3402-40ee-b509-51c54f8af982-kube-api-access-62jbp\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215007 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215028 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/945666f1-2703-4fa7-9063-fd49b7189341-serving-cert\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215050 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215071 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e0a197c-3a94-4426-a492-d541a18194a0-audit-dir\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215092 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215113 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-stats-auth\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215134 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/caab36f1-979c-459d-b742-941e13dc765e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215157 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glzq8\" (UniqueName: \"kubernetes.io/projected/15c836ec-fd1e-47ee-bfc7-90c419082e8b-kube-api-access-glzq8\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215172 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215180 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dghq\" (UniqueName: \"kubernetes.io/projected/8c6d92f4-c514-48e4-b1da-61546f07d171-kube-api-access-2dghq\") pod \"control-plane-machine-set-operator-78cbb6b69f-hln48\" (UID: \"8c6d92f4-c514-48e4-b1da-61546f07d171\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215294 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-encryption-config\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215354 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-oauth-config\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215389 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-machine-approver-tls\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215478 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-default-certificate\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215526 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215529 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tpr6j"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215548 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/733daf88-a275-4be5-ace0-0d12edaf1b87-config\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215575 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt6bn\" (UniqueName: \"kubernetes.io/projected/5461c713-e58f-46e8-91f7-5d6546011e27-kube-api-access-dt6bn\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.215598 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-serving-cert\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216408 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/945666f1-2703-4fa7-9063-fd49b7189341-config\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216453 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216623 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216648 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5461c713-e58f-46e8-91f7-5d6546011e27-node-pullsecrets\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216708 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/418bec94-3753-4d76-b2a8-24b04f39c820-images\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216737 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-client\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216759 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-config\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216781 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216847 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216887 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-serving-cert\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216911 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216935 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g2b2\" (UniqueName: \"kubernetes.io/projected/418bec94-3753-4d76-b2a8-24b04f39c820-kube-api-access-6g2b2\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216940 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-config\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216957 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.216981 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217003 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-config\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217012 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5461c713-e58f-46e8-91f7-5d6546011e27-audit-dir\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217037 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217105 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f60bac9-dca3-4eb8-b163-8146b084e74d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217131 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-serving-cert\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217157 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h49q\" (UniqueName: \"kubernetes.io/projected/918fce5c-2c8a-46a8-9c16-3911fb36faa3-kube-api-access-6h49q\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217178 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-encryption-config\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217199 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-config\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217222 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-etcd-client\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217375 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-policies\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217694 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99f7833f-00a4-45fe-80ba-6c22404c49e4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217729 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsjmz\" (UniqueName: \"kubernetes.io/projected/2037c5ee-5855-4520-af01-000ac44338bb-kube-api-access-lsjmz\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.217777 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-client-ca\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.218252 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-image-import-ca\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.218867 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-config\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.219508 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.219661 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-service-ca\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.219861 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-config\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.221434 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.222199 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.222868 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.222886 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-audit\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.222986 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223020 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-client-ca\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223051 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wxqx\" (UniqueName: \"kubernetes.io/projected/01e125dd-87e2-4362-842e-849288698994-kube-api-access-8wxqx\") pod \"auto-csr-approver-29556462-xxthz\" (UID: \"01e125dd-87e2-4362-842e-849288698994\") " pod="openshift-infra/auto-csr-approver-29556462-xxthz" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223075 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c836ec-fd1e-47ee-bfc7-90c419082e8b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223124 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-config\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223152 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99f7833f-00a4-45fe-80ba-6c22404c49e4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223179 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99f7833f-00a4-45fe-80ba-6c22404c49e4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223204 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223227 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418bec94-3753-4d76-b2a8-24b04f39c820-config\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223290 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-trusted-ca-bundle\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223320 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223344 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/945666f1-2703-4fa7-9063-fd49b7189341-trusted-ca\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223368 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg92p\" (UniqueName: \"kubernetes.io/projected/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-kube-api-access-rg92p\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223395 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-profile-collector-cert\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223420 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-dir\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223443 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdzcp\" (UniqueName: \"kubernetes.io/projected/945666f1-2703-4fa7-9063-fd49b7189341-kube-api-access-gdzcp\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223472 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223477 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-audit\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223498 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1fe6fb32-9851-48de-88f6-94701fcf7010-service-ca-bundle\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223554 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-config\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223581 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-srv-cert\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223607 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-auth-proxy-config\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223635 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n6t6\" (UniqueName: \"kubernetes.io/projected/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-kube-api-access-4n6t6\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223659 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xflh\" (UniqueName: \"kubernetes.io/projected/90d11aee-636b-4d59-bb06-017a3adafe81-kube-api-access-8xflh\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223685 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918fce5c-2c8a-46a8-9c16-3911fb36faa3-serving-cert\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223708 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-ca\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223730 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-serving-cert\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223752 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caab36f1-979c-459d-b742-941e13dc765e-serving-cert\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223778 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a8ea6bd-3402-40ee-b509-51c54f8af982-serving-cert\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223800 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-etcd-serving-ca\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223794 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/733daf88-a275-4be5-ace0-0d12edaf1b87-config\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223829 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-service-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223947 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-serving-cert\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.223997 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c59f6\" (UniqueName: \"kubernetes.io/projected/62fdbab2-2be8-42cc-a936-a87ffd085a0e-kube-api-access-c59f6\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224021 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c6d92f4-c514-48e4-b1da-61546f07d171-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hln48\" (UID: \"8c6d92f4-c514-48e4-b1da-61546f07d171\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224051 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-service-ca\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224069 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224112 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr74m\" (UniqueName: \"kubernetes.io/projected/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-kube-api-access-sr74m\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224132 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-srv-cert\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224148 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-etcd-client\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224163 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224195 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72bdr\" (UniqueName: \"kubernetes.io/projected/99f7833f-00a4-45fe-80ba-6c22404c49e4-kube-api-access-72bdr\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224217 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224236 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15c836ec-fd1e-47ee-bfc7-90c419082e8b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224274 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcmxf\" (UniqueName: \"kubernetes.io/projected/6f60bac9-dca3-4eb8-b163-8146b084e74d-kube-api-access-dcmxf\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224473 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-service-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.224501 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5461c713-e58f-46e8-91f7-5d6546011e27-node-pullsecrets\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.225264 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e0a197c-3a94-4426-a492-d541a18194a0-audit-dir\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.225793 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-75kbq"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.246306 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.239337 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.239979 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/945666f1-2703-4fa7-9063-fd49b7189341-serving-cert\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.240138 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.240157 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-config\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.246539 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-trusted-ca-bundle\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.241947 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-oauth-serving-cert\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.242029 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-client-ca\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.242139 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.242422 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a8ea6bd-3402-40ee-b509-51c54f8af982-config\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.242576 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.242724 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99f7833f-00a4-45fe-80ba-6c22404c49e4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.242946 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.246950 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-metrics-certs\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.243607 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/418bec94-3753-4d76-b2a8-24b04f39c820-images\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.244523 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418bec94-3753-4d76-b2a8-24b04f39c820-config\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.245197 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-serving-cert\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.245474 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.245520 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-client\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.245563 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.245914 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-service-ca\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.245997 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-serving-cert\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.246190 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-encryption-config\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.246202 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-oauth-config\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.247176 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/733daf88-a275-4be5-ace0-0d12edaf1b87-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.246312 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-default-certificate\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.246332 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.239211 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.240814 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-audit-policies\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.247671 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-serving-cert\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.248054 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/418bec94-3753-4d76-b2a8-24b04f39c820-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.248183 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-config\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.248747 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.249516 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.250178 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-client-ca\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.244570 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.250663 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.251907 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-etcd-client\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.254615 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m2gxf"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.255150 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/945666f1-2703-4fa7-9063-fd49b7189341-config\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.256336 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.256686 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918fce5c-2c8a-46a8-9c16-3911fb36faa3-serving-cert\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.256921 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a8ea6bd-3402-40ee-b509-51c54f8af982-serving-cert\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.256969 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.257147 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/945666f1-2703-4fa7-9063-fd49b7189341-trusted-ca\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.257501 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1fe6fb32-9851-48de-88f6-94701fcf7010-service-ca-bundle\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.257549 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/caab36f1-979c-459d-b742-941e13dc765e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.257728 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-dir\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.258409 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-config\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.243363 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e0a197c-3a94-4426-a492-d541a18194a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.259108 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5461c713-e58f-46e8-91f7-5d6546011e27-etcd-serving-ca\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.256650 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/918fce5c-2c8a-46a8-9c16-3911fb36faa3-etcd-ca\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.259618 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.259990 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2bszj"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.260780 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.262988 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5461c713-e58f-46e8-91f7-5d6546011e27-serving-cert\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.263043 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-etcd-client\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.263438 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1fe6fb32-9851-48de-88f6-94701fcf7010-stats-auth\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.263899 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dfkkx"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.264481 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-encryption-config\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.264913 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e0a197c-3a94-4426-a492-d541a18194a0-serving-cert\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.266366 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caab36f1-979c-459d-b742-941e13dc765e-serving-cert\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.268058 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6k6wd"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.268329 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99f7833f-00a4-45fe-80ba-6c22404c49e4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.269521 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.272066 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.273458 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-zk5kl"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.274426 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zk5kl" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.274821 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zk5kl"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.276046 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-vnfvs"] Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.276814 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.281507 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.300702 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.320746 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325046 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsjmz\" (UniqueName: \"kubernetes.io/projected/2037c5ee-5855-4520-af01-000ac44338bb-kube-api-access-lsjmz\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325161 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wxqx\" (UniqueName: \"kubernetes.io/projected/01e125dd-87e2-4362-842e-849288698994-kube-api-access-8wxqx\") pod \"auto-csr-approver-29556462-xxthz\" (UID: \"01e125dd-87e2-4362-842e-849288698994\") " pod="openshift-infra/auto-csr-approver-29556462-xxthz" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325256 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c836ec-fd1e-47ee-bfc7-90c419082e8b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325351 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-profile-collector-cert\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325434 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg92p\" (UniqueName: \"kubernetes.io/projected/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-kube-api-access-rg92p\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325519 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-srv-cert\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325591 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-auth-proxy-config\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325669 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xflh\" (UniqueName: \"kubernetes.io/projected/90d11aee-636b-4d59-bb06-017a3adafe81-kube-api-access-8xflh\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325758 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c6d92f4-c514-48e4-b1da-61546f07d171-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hln48\" (UID: \"8c6d92f4-c514-48e4-b1da-61546f07d171\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325853 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-srv-cert\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.325936 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15c836ec-fd1e-47ee-bfc7-90c419082e8b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326018 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcmxf\" (UniqueName: \"kubernetes.io/projected/6f60bac9-dca3-4eb8-b163-8146b084e74d-kube-api-access-dcmxf\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326098 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd2j2\" (UniqueName: \"kubernetes.io/projected/300d0a76-b532-4124-af3d-f89cf7a51d3f-kube-api-access-vd2j2\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326186 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f60bac9-dca3-4eb8-b163-8146b084e74d-proxy-tls\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326302 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2037c5ee-5855-4520-af01-000ac44338bb-signing-cabundle\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326421 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2037c5ee-5855-4520-af01-000ac44338bb-signing-key\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326490 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glzq8\" (UniqueName: \"kubernetes.io/projected/15c836ec-fd1e-47ee-bfc7-90c419082e8b-kube-api-access-glzq8\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326566 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dghq\" (UniqueName: \"kubernetes.io/projected/8c6d92f4-c514-48e4-b1da-61546f07d171-kube-api-access-2dghq\") pod \"control-plane-machine-set-operator-78cbb6b69f-hln48\" (UID: \"8c6d92f4-c514-48e4-b1da-61546f07d171\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326645 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-machine-approver-tls\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326808 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-config\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.326921 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f60bac9-dca3-4eb8-b163-8146b084e74d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.327997 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f60bac9-dca3-4eb8-b163-8146b084e74d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.341504 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.361237 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.381699 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.401049 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.420877 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.441369 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.461382 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.481205 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.501683 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.521319 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.529629 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-machine-approver-tls\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.541722 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.547411 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-auth-proxy-config\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.562058 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.568084 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-config\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.581811 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.601959 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.621699 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.642075 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.660787 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.680967 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.700637 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.723534 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.751183 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.761881 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.781610 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.801421 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.822001 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.842336 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.861793 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.881843 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.901547 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.907092 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c836ec-fd1e-47ee-bfc7-90c419082e8b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.921524 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.929481 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f60bac9-dca3-4eb8-b163-8146b084e74d-proxy-tls\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.941617 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.961679 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.968756 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15c836ec-fd1e-47ee-bfc7-90c419082e8b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:43 crc kubenswrapper[4876]: I0313 07:42:43.981538 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.002658 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.022138 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.041041 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.061595 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.081741 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.099655 4876 request.go:700] Waited for 1.009633351s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.101705 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.121773 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.129376 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c6d92f4-c514-48e4-b1da-61546f07d171-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hln48\" (UID: \"8c6d92f4-c514-48e4-b1da-61546f07d171\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.141414 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.161980 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.189722 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.201230 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.225823 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.242019 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.281913 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.301262 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.322469 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.325750 4876 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.325797 4876 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.325846 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-profile-collector-cert podName:90d11aee-636b-4d59-bb06-017a3adafe81 nodeName:}" failed. No retries permitted until 2026-03-13 07:42:44.825819635 +0000 UTC m=+224.496598637 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-profile-collector-cert") pod "catalog-operator-68c6474976-vwq9g" (UID: "90d11aee-636b-4d59-bb06-017a3adafe81") : failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.325948 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-srv-cert podName:300d0a76-b532-4124-af3d-f89cf7a51d3f nodeName:}" failed. No retries permitted until 2026-03-13 07:42:44.825914348 +0000 UTC m=+224.496693380 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-srv-cert") pod "olm-operator-6b444d44fb-z8697" (UID: "300d0a76-b532-4124-af3d-f89cf7a51d3f") : failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.325977 4876 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.326279 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-srv-cert podName:90d11aee-636b-4d59-bb06-017a3adafe81 nodeName:}" failed. No retries permitted until 2026-03-13 07:42:44.826057122 +0000 UTC m=+224.496836164 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-srv-cert") pod "catalog-operator-68c6474976-vwq9g" (UID: "90d11aee-636b-4d59-bb06-017a3adafe81") : failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.326946 4876 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.326981 4876 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.327024 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2037c5ee-5855-4520-af01-000ac44338bb-signing-key podName:2037c5ee-5855-4520-af01-000ac44338bb nodeName:}" failed. No retries permitted until 2026-03-13 07:42:44.827004709 +0000 UTC m=+224.497783731 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/2037c5ee-5855-4520-af01-000ac44338bb-signing-key") pod "service-ca-9c57cc56f-6k6wd" (UID: "2037c5ee-5855-4520-af01-000ac44338bb") : failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.327052 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-profile-collector-cert podName:300d0a76-b532-4124-af3d-f89cf7a51d3f nodeName:}" failed. No retries permitted until 2026-03-13 07:42:44.82703934 +0000 UTC m=+224.497818362 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-profile-collector-cert") pod "olm-operator-6b444d44fb-z8697" (UID: "300d0a76-b532-4124-af3d-f89cf7a51d3f") : failed to sync secret cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.327060 4876 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: E0313 07:42:44.327107 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2037c5ee-5855-4520-af01-000ac44338bb-signing-cabundle podName:2037c5ee-5855-4520-af01-000ac44338bb nodeName:}" failed. No retries permitted until 2026-03-13 07:42:44.827090651 +0000 UTC m=+224.497869703 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/2037c5ee-5855-4520-af01-000ac44338bb-signing-cabundle") pod "service-ca-9c57cc56f-6k6wd" (UID: "2037c5ee-5855-4520-af01-000ac44338bb") : failed to sync configmap cache: timed out waiting for the condition Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.341126 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.361337 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.381037 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.400857 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.421373 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.441405 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.461588 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.481855 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.501136 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.522091 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.541897 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.562284 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.581619 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.601036 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.621308 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.641361 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.661210 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.681873 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.702377 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.721935 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.742165 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.762482 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.782859 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.823580 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.842344 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.846211 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-profile-collector-cert\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.846377 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-srv-cert\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.846529 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-srv-cert\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.846789 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2037c5ee-5855-4520-af01-000ac44338bb-signing-cabundle\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.847550 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2037c5ee-5855-4520-af01-000ac44338bb-signing-key\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.847742 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.850037 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2037c5ee-5855-4520-af01-000ac44338bb-signing-cabundle\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.852137 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2037c5ee-5855-4520-af01-000ac44338bb-signing-key\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.853991 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.854080 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/300d0a76-b532-4124-af3d-f89cf7a51d3f-srv-cert\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.854214 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-profile-collector-cert\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.854443 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90d11aee-636b-4d59-bb06-017a3adafe81-srv-cert\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.862605 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.882096 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.901180 4876 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.920647 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.960633 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62jbp\" (UniqueName: \"kubernetes.io/projected/0a8ea6bd-3402-40ee-b509-51c54f8af982-kube-api-access-62jbp\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"0a8ea6bd-3402-40ee-b509-51c54f8af982\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:44 crc kubenswrapper[4876]: I0313 07:42:44.978185 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/733daf88-a275-4be5-ace0-0d12edaf1b87-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hlwjw\" (UID: \"733daf88-a275-4be5-ace0-0d12edaf1b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.003286 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf48g\" (UniqueName: \"kubernetes.io/projected/6b92045d-38c7-4dff-be0a-78449d870bcb-kube-api-access-mf48g\") pod \"oauth-openshift-558db77b4-xvfbt\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.016114 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drh79\" (UniqueName: \"kubernetes.io/projected/5e0a197c-3a94-4426-a492-d541a18194a0-kube-api-access-drh79\") pod \"apiserver-7bbb656c7d-xttrq\" (UID: \"5e0a197c-3a94-4426-a492-d541a18194a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.036338 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szj2c\" (UniqueName: \"kubernetes.io/projected/1fe6fb32-9851-48de-88f6-94701fcf7010-kube-api-access-szj2c\") pod \"router-default-5444994796-cd4ct\" (UID: \"1fe6fb32-9851-48de-88f6-94701fcf7010\") " pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.062422 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97r5d\" (UniqueName: \"kubernetes.io/projected/caab36f1-979c-459d-b742-941e13dc765e-kube-api-access-97r5d\") pod \"openshift-config-operator-7777fb866f-7f5nt\" (UID: \"caab36f1-979c-459d-b742-941e13dc765e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.078366 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cttzt\" (UniqueName: \"kubernetes.io/projected/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-kube-api-access-cttzt\") pod \"controller-manager-879f6c89f-nkhkk\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.096942 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkcd2\" (UniqueName: \"kubernetes.io/projected/ee3187e1-6603-4cd1-a134-6437fae1ca4a-kube-api-access-fkcd2\") pod \"downloads-7954f5f757-6cnvw\" (UID: \"ee3187e1-6603-4cd1-a134-6437fae1ca4a\") " pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.100154 4876 request.go:700] Waited for 1.878063681s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver/serviceaccounts/openshift-apiserver-sa/token Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.115475 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt6bn\" (UniqueName: \"kubernetes.io/projected/5461c713-e58f-46e8-91f7-5d6546011e27-kube-api-access-dt6bn\") pod \"apiserver-76f77b778f-6qqdn\" (UID: \"5461c713-e58f-46e8-91f7-5d6546011e27\") " pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.129505 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.138009 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c59f6\" (UniqueName: \"kubernetes.io/projected/62fdbab2-2be8-42cc-a936-a87ffd085a0e-kube-api-access-c59f6\") pod \"console-f9d7485db-g65sf\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.143410 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.165944 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g2b2\" (UniqueName: \"kubernetes.io/projected/418bec94-3753-4d76-b2a8-24b04f39c820-kube-api-access-6g2b2\") pod \"machine-api-operator-5694c8668f-xvwj2\" (UID: \"418bec94-3753-4d76-b2a8-24b04f39c820\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.179393 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72bdr\" (UniqueName: \"kubernetes.io/projected/99f7833f-00a4-45fe-80ba-6c22404c49e4-kube-api-access-72bdr\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.184073 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.198701 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pnjqr\" (UID: \"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.212668 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.234133 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.238158 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h49q\" (UniqueName: \"kubernetes.io/projected/918fce5c-2c8a-46a8-9c16-3911fb36faa3-kube-api-access-6h49q\") pod \"etcd-operator-b45778765-lqr7r\" (UID: \"918fce5c-2c8a-46a8-9c16-3911fb36faa3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.238832 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.247997 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99f7833f-00a4-45fe-80ba-6c22404c49e4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fbjjp\" (UID: \"99f7833f-00a4-45fe-80ba-6c22404c49e4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.256391 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.263401 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr74m\" (UniqueName: \"kubernetes.io/projected/29aded1a-dd1a-4ce1-b441-b3b897e4a06d-kube-api-access-sr74m\") pod \"openshift-apiserver-operator-796bbdcf4f-mb8k7\" (UID: \"29aded1a-dd1a-4ce1-b441-b3b897e4a06d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.266545 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.277561 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n6t6\" (UniqueName: \"kubernetes.io/projected/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-kube-api-access-4n6t6\") pod \"route-controller-manager-6576b87f9c-jdqxn\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.281343 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.287565 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.297170 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.301120 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.301199 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.313941 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdzcp\" (UniqueName: \"kubernetes.io/projected/945666f1-2703-4fa7-9063-fd49b7189341-kube-api-access-gdzcp\") pod \"console-operator-58897d9998-5rhkc\" (UID: \"945666f1-2703-4fa7-9063-fd49b7189341\") " pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.321844 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.337591 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6qqdn"] Mar 13 07:42:45 crc kubenswrapper[4876]: W0313 07:42:45.340192 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fe6fb32_9851_48de_88f6_94701fcf7010.slice/crio-8c12e4f91be7a690b5a0df7b51d7e79a0f80a93d71b86b7a6164d11b755bba45 WatchSource:0}: Error finding container 8c12e4f91be7a690b5a0df7b51d7e79a0f80a93d71b86b7a6164d11b755bba45: Status 404 returned error can't find the container with id 8c12e4f91be7a690b5a0df7b51d7e79a0f80a93d71b86b7a6164d11b755bba45 Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.340735 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.355358 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.361516 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.372619 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" Mar 13 07:42:45 crc kubenswrapper[4876]: W0313 07:42:45.374071 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5461c713_e58f_46e8_91f7_5d6546011e27.slice/crio-cc636e8704c8c1039589602a55bd37772fd44761db79934b7aa5522efc495aa2 WatchSource:0}: Error finding container cc636e8704c8c1039589602a55bd37772fd44761db79934b7aa5522efc495aa2: Status 404 returned error can't find the container with id cc636e8704c8c1039589602a55bd37772fd44761db79934b7aa5522efc495aa2 Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.387874 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.389121 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.397460 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.403530 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.428426 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.503776 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wxqx\" (UniqueName: \"kubernetes.io/projected/01e125dd-87e2-4362-842e-849288698994-kube-api-access-8wxqx\") pod \"auto-csr-approver-29556462-xxthz\" (UID: \"01e125dd-87e2-4362-842e-849288698994\") " pod="openshift-infra/auto-csr-approver-29556462-xxthz" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.510770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsjmz\" (UniqueName: \"kubernetes.io/projected/2037c5ee-5855-4520-af01-000ac44338bb-kube-api-access-lsjmz\") pod \"service-ca-9c57cc56f-6k6wd\" (UID: \"2037c5ee-5855-4520-af01-000ac44338bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.514710 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg92p\" (UniqueName: \"kubernetes.io/projected/34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4-kube-api-access-rg92p\") pod \"machine-approver-56656f9798-dnr6c\" (UID: \"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.515410 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.526664 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xflh\" (UniqueName: \"kubernetes.io/projected/90d11aee-636b-4d59-bb06-017a3adafe81-kube-api-access-8xflh\") pod \"catalog-operator-68c6474976-vwq9g\" (UID: \"90d11aee-636b-4d59-bb06-017a3adafe81\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.546144 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcmxf\" (UniqueName: \"kubernetes.io/projected/6f60bac9-dca3-4eb8-b163-8146b084e74d-kube-api-access-dcmxf\") pod \"machine-config-controller-84d6567774-jn5d4\" (UID: \"6f60bac9-dca3-4eb8-b163-8146b084e74d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.558127 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd2j2\" (UniqueName: \"kubernetes.io/projected/300d0a76-b532-4124-af3d-f89cf7a51d3f-kube-api-access-vd2j2\") pod \"olm-operator-6b444d44fb-z8697\" (UID: \"300d0a76-b532-4124-af3d-f89cf7a51d3f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.574138 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.580709 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glzq8\" (UniqueName: \"kubernetes.io/projected/15c836ec-fd1e-47ee-bfc7-90c419082e8b-kube-api-access-glzq8\") pod \"kube-storage-version-migrator-operator-b67b599dd-dlvtf\" (UID: \"15c836ec-fd1e-47ee-bfc7-90c419082e8b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.583557 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq"] Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.602038 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dghq\" (UniqueName: \"kubernetes.io/projected/8c6d92f4-c514-48e4-b1da-61546f07d171-kube-api-access-2dghq\") pod \"control-plane-machine-set-operator-78cbb6b69f-hln48\" (UID: \"8c6d92f4-c514-48e4-b1da-61546f07d171\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.663834 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.663873 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.663895 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f938364d-5a03-47e8-b863-edf8d217770a-apiservice-cert\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.663915 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-certificates\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.663961 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k6gg\" (UniqueName: \"kubernetes.io/projected/b82a006d-154c-4e07-b683-c65ca4863459-kube-api-access-7k6gg\") pod \"migrator-59844c95c7-zktvd\" (UID: \"b82a006d-154c-4e07-b683-c65ca4863459\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.663983 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-bound-sa-token\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664053 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-serving-cert\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664098 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f938364d-5a03-47e8-b863-edf8d217770a-tmpfs\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664123 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7984a28c-6710-47b1-9c43-7e2c39dda944-images\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664149 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c59f72f6-172a-4b6e-a367-470966fdb1c1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664185 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw8lv\" (UniqueName: \"kubernetes.io/projected/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-kube-api-access-hw8lv\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664222 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7984a28c-6710-47b1-9c43-7e2c39dda944-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664257 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664280 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-tls\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664334 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ldtm\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-kube-api-access-5ldtm\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664373 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664398 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664447 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc8b898-1330-4381-b780-ee41a9514f05-config-volume\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664469 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664491 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/906277a0-614e-442f-873f-4d1d67e4ae5d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664534 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efc8b898-1330-4381-b780-ee41a9514f05-secret-volume\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664556 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c59f72f6-172a-4b6e-a367-470966fdb1c1-trusted-ca\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664576 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qppc7\" (UniqueName: \"kubernetes.io/projected/f089ab79-2e29-4074-a8dc-21b0af79d3f7-kube-api-access-qppc7\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664597 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7984a28c-6710-47b1-9c43-7e2c39dda944-proxy-tls\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664618 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r655p\" (UniqueName: \"kubernetes.io/projected/7984a28c-6710-47b1-9c43-7e2c39dda944-kube-api-access-r655p\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664641 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfhhk\" (UniqueName: \"kubernetes.io/projected/0445a35a-4d10-444b-99ba-b0ced7f82f09-kube-api-access-rfhhk\") pod \"multus-admission-controller-857f4d67dd-lw2ss\" (UID: \"0445a35a-4d10-444b-99ba-b0ced7f82f09\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664663 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkns5\" (UniqueName: \"kubernetes.io/projected/a41ab691-ce98-4128-a110-6fb02708545e-kube-api-access-kkns5\") pod \"cluster-samples-operator-665b6dd947-4pkp6\" (UID: \"a41ab691-ce98-4128-a110-6fb02708545e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-trusted-ca\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664735 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsk25\" (UniqueName: \"kubernetes.io/projected/ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2-kube-api-access-rsk25\") pod \"package-server-manager-789f6589d5-wbv9l\" (UID: \"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664761 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-config\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664796 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0445a35a-4d10-444b-99ba-b0ced7f82f09-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lw2ss\" (UID: \"0445a35a-4d10-444b-99ba-b0ced7f82f09\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664815 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c59f72f6-172a-4b6e-a367-470966fdb1c1-metrics-tls\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664848 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb28d\" (UniqueName: \"kubernetes.io/projected/f938364d-5a03-47e8-b863-edf8d217770a-kube-api-access-sb28d\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664867 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f-metrics-tls\") pod \"dns-operator-744455d44c-tpr6j\" (UID: \"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f\") " pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664921 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f938364d-5a03-47e8-b863-edf8d217770a-webhook-cert\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664943 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d26n\" (UniqueName: \"kubernetes.io/projected/906277a0-614e-442f-873f-4d1d67e4ae5d-kube-api-access-4d26n\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664977 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wbv9l\" (UID: \"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.664997 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g8z6\" (UniqueName: \"kubernetes.io/projected/e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f-kube-api-access-2g8z6\") pod \"dns-operator-744455d44c-tpr6j\" (UID: \"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f\") " pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.665017 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpvw6\" (UniqueName: \"kubernetes.io/projected/c59f72f6-172a-4b6e-a367-470966fdb1c1-kube-api-access-hpvw6\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.665096 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnljr\" (UniqueName: \"kubernetes.io/projected/efc8b898-1330-4381-b780-ee41a9514f05-kube-api-access-bnljr\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.665117 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.665139 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a41ab691-ce98-4128-a110-6fb02708545e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4pkp6\" (UID: \"a41ab691-ce98-4128-a110-6fb02708545e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.665191 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/906277a0-614e-442f-873f-4d1d67e4ae5d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.665241 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: E0313 07:42:45.674683 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.174659071 +0000 UTC m=+225.845438083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.685171 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.727764 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.736523 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.750974 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.767878 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768098 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768121 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f938364d-5a03-47e8-b863-edf8d217770a-apiservice-cert\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768143 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-mountpoint-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768160 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-certificates\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768215 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k6gg\" (UniqueName: \"kubernetes.io/projected/b82a006d-154c-4e07-b683-c65ca4863459-kube-api-access-7k6gg\") pod \"migrator-59844c95c7-zktvd\" (UID: \"b82a006d-154c-4e07-b683-c65ca4863459\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768255 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-bound-sa-token\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768289 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-socket-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768315 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-serving-cert\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768345 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/38a6f1e3-c474-4f1d-961a-e147adfcbe72-certs\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768359 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-plugins-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768378 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f938364d-5a03-47e8-b863-edf8d217770a-tmpfs\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768394 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qbl6\" (UniqueName: \"kubernetes.io/projected/c0d175eb-6c81-43c8-97d4-ed36ac824321-kube-api-access-6qbl6\") pod \"ingress-canary-zk5kl\" (UID: \"c0d175eb-6c81-43c8-97d4-ed36ac824321\") " pod="openshift-ingress-canary/ingress-canary-zk5kl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768418 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7984a28c-6710-47b1-9c43-7e2c39dda944-images\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768441 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c59f72f6-172a-4b6e-a367-470966fdb1c1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768460 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw8lv\" (UniqueName: \"kubernetes.io/projected/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-kube-api-access-hw8lv\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768475 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-csi-data-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768493 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7984a28c-6710-47b1-9c43-7e2c39dda944-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768508 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768540 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-tls\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768565 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ldtm\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-kube-api-access-5ldtm\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768580 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768595 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768609 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-registration-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768635 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc8b898-1330-4381-b780-ee41a9514f05-config-volume\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768652 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768666 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njsxl\" (UniqueName: \"kubernetes.io/projected/38a6f1e3-c474-4f1d-961a-e147adfcbe72-kube-api-access-njsxl\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768684 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/906277a0-614e-442f-873f-4d1d67e4ae5d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768699 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efc8b898-1330-4381-b780-ee41a9514f05-secret-volume\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768715 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54fwt\" (UniqueName: \"kubernetes.io/projected/aa7285c8-4679-495d-afca-6d15a19cfa4c-kube-api-access-54fwt\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768751 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c59f72f6-172a-4b6e-a367-470966fdb1c1-trusted-ca\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768767 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qppc7\" (UniqueName: \"kubernetes.io/projected/f089ab79-2e29-4074-a8dc-21b0af79d3f7-kube-api-access-qppc7\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768784 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7984a28c-6710-47b1-9c43-7e2c39dda944-proxy-tls\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768803 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r655p\" (UniqueName: \"kubernetes.io/projected/7984a28c-6710-47b1-9c43-7e2c39dda944-kube-api-access-r655p\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768820 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfhhk\" (UniqueName: \"kubernetes.io/projected/0445a35a-4d10-444b-99ba-b0ced7f82f09-kube-api-access-rfhhk\") pod \"multus-admission-controller-857f4d67dd-lw2ss\" (UID: \"0445a35a-4d10-444b-99ba-b0ced7f82f09\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768836 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkns5\" (UniqueName: \"kubernetes.io/projected/a41ab691-ce98-4128-a110-6fb02708545e-kube-api-access-kkns5\") pod \"cluster-samples-operator-665b6dd947-4pkp6\" (UID: \"a41ab691-ce98-4128-a110-6fb02708545e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768872 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/38a6f1e3-c474-4f1d-961a-e147adfcbe72-node-bootstrap-token\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768889 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxbvz\" (UniqueName: \"kubernetes.io/projected/ed1931f4-3008-4ee4-8295-0c0c99669cf1-kube-api-access-jxbvz\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768914 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-trusted-ca\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768938 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsk25\" (UniqueName: \"kubernetes.io/projected/ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2-kube-api-access-rsk25\") pod \"package-server-manager-789f6589d5-wbv9l\" (UID: \"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768953 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-config\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768976 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0445a35a-4d10-444b-99ba-b0ced7f82f09-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lw2ss\" (UID: \"0445a35a-4d10-444b-99ba-b0ced7f82f09\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.768990 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c59f72f6-172a-4b6e-a367-470966fdb1c1-metrics-tls\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769006 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb28d\" (UniqueName: \"kubernetes.io/projected/f938364d-5a03-47e8-b863-edf8d217770a-kube-api-access-sb28d\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769024 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f-metrics-tls\") pod \"dns-operator-744455d44c-tpr6j\" (UID: \"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f\") " pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769038 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0d175eb-6c81-43c8-97d4-ed36ac824321-cert\") pod \"ingress-canary-zk5kl\" (UID: \"c0d175eb-6c81-43c8-97d4-ed36ac824321\") " pod="openshift-ingress-canary/ingress-canary-zk5kl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769070 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f938364d-5a03-47e8-b863-edf8d217770a-webhook-cert\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769095 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d26n\" (UniqueName: \"kubernetes.io/projected/906277a0-614e-442f-873f-4d1d67e4ae5d-kube-api-access-4d26n\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769111 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed1931f4-3008-4ee4-8295-0c0c99669cf1-metrics-tls\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769126 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wbv9l\" (UID: \"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769142 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g8z6\" (UniqueName: \"kubernetes.io/projected/e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f-kube-api-access-2g8z6\") pod \"dns-operator-744455d44c-tpr6j\" (UID: \"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f\") " pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769184 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpvw6\" (UniqueName: \"kubernetes.io/projected/c59f72f6-172a-4b6e-a367-470966fdb1c1-kube-api-access-hpvw6\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769207 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnljr\" (UniqueName: \"kubernetes.io/projected/efc8b898-1330-4381-b780-ee41a9514f05-kube-api-access-bnljr\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769222 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769264 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a41ab691-ce98-4128-a110-6fb02708545e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4pkp6\" (UID: \"a41ab691-ce98-4128-a110-6fb02708545e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769292 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/906277a0-614e-442f-873f-4d1d67e4ae5d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769307 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed1931f4-3008-4ee4-8295-0c0c99669cf1-config-volume\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.769342 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: E0313 07:42:45.769524 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.269509307 +0000 UTC m=+225.940288279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.775355 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f938364d-5a03-47e8-b863-edf8d217770a-apiservice-cert\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.775811 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.776480 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-certificates\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.781623 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7984a28c-6710-47b1-9c43-7e2c39dda944-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.781647 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.782149 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7984a28c-6710-47b1-9c43-7e2c39dda944-proxy-tls\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.782219 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc8b898-1330-4381-b780-ee41a9514f05-config-volume\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.783232 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.784206 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f938364d-5a03-47e8-b863-edf8d217770a-tmpfs\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.785256 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c59f72f6-172a-4b6e-a367-470966fdb1c1-trusted-ca\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.793067 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/906277a0-614e-442f-873f-4d1d67e4ae5d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.795232 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-config\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.801545 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556462-xxthz" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.807813 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efc8b898-1330-4381-b780-ee41a9514f05-secret-volume\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.807959 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-tls\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.807963 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7984a28c-6710-47b1-9c43-7e2c39dda944-images\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.808407 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.808490 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c59f72f6-172a-4b6e-a367-470966fdb1c1-metrics-tls\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.808841 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.808861 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f-metrics-tls\") pod \"dns-operator-744455d44c-tpr6j\" (UID: \"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f\") " pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.808925 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a41ab691-ce98-4128-a110-6fb02708545e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4pkp6\" (UID: \"a41ab691-ce98-4128-a110-6fb02708545e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.808937 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0445a35a-4d10-444b-99ba-b0ced7f82f09-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lw2ss\" (UID: \"0445a35a-4d10-444b-99ba-b0ced7f82f09\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.809100 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f938364d-5a03-47e8-b863-edf8d217770a-webhook-cert\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.809413 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-serving-cert\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.809443 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.809881 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-trusted-ca\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.809900 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.810740 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wbv9l\" (UID: \"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.812150 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xvfbt"] Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.817641 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/906277a0-614e-442f-873f-4d1d67e4ae5d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.828980 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qppc7\" (UniqueName: \"kubernetes.io/projected/f089ab79-2e29-4074-a8dc-21b0af79d3f7-kube-api-access-qppc7\") pod \"marketplace-operator-79b997595-75kbq\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.831343 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.844323 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k6gg\" (UniqueName: \"kubernetes.io/projected/b82a006d-154c-4e07-b683-c65ca4863459-kube-api-access-7k6gg\") pod \"migrator-59844c95c7-zktvd\" (UID: \"b82a006d-154c-4e07-b683-c65ca4863459\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.860436 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-bound-sa-token\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.870613 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed1931f4-3008-4ee4-8295-0c0c99669cf1-metrics-tls\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871121 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed1931f4-3008-4ee4-8295-0c0c99669cf1-config-volume\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871159 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-mountpoint-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871179 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871213 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-socket-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871280 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/38a6f1e3-c474-4f1d-961a-e147adfcbe72-certs\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871305 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-plugins-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871327 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qbl6\" (UniqueName: \"kubernetes.io/projected/c0d175eb-6c81-43c8-97d4-ed36ac824321-kube-api-access-6qbl6\") pod \"ingress-canary-zk5kl\" (UID: \"c0d175eb-6c81-43c8-97d4-ed36ac824321\") " pod="openshift-ingress-canary/ingress-canary-zk5kl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871363 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-csi-data-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871400 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-registration-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871424 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njsxl\" (UniqueName: \"kubernetes.io/projected/38a6f1e3-c474-4f1d-961a-e147adfcbe72-kube-api-access-njsxl\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871443 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54fwt\" (UniqueName: \"kubernetes.io/projected/aa7285c8-4679-495d-afca-6d15a19cfa4c-kube-api-access-54fwt\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871477 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/38a6f1e3-c474-4f1d-961a-e147adfcbe72-node-bootstrap-token\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871497 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxbvz\" (UniqueName: \"kubernetes.io/projected/ed1931f4-3008-4ee4-8295-0c0c99669cf1-kube-api-access-jxbvz\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.871522 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0d175eb-6c81-43c8-97d4-ed36ac824321-cert\") pod \"ingress-canary-zk5kl\" (UID: \"c0d175eb-6c81-43c8-97d4-ed36ac824321\") " pod="openshift-ingress-canary/ingress-canary-zk5kl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.872138 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-csi-data-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.872326 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-registration-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.872467 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-socket-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.872498 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed1931f4-3008-4ee4-8295-0c0c99669cf1-config-volume\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:45 crc kubenswrapper[4876]: E0313 07:42:45.872917 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.37290229 +0000 UTC m=+226.043681272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.872529 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-mountpoint-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.873400 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa7285c8-4679-495d-afca-6d15a19cfa4c-plugins-dir\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.879435 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/38a6f1e3-c474-4f1d-961a-e147adfcbe72-certs\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.880597 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed1931f4-3008-4ee4-8295-0c0c99669cf1-metrics-tls\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.881158 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/38a6f1e3-c474-4f1d-961a-e147adfcbe72-node-bootstrap-token\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.882061 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r655p\" (UniqueName: \"kubernetes.io/projected/7984a28c-6710-47b1-9c43-7e2c39dda944-kube-api-access-r655p\") pod \"machine-config-operator-74547568cd-xkfqg\" (UID: \"7984a28c-6710-47b1-9c43-7e2c39dda944\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.889947 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0d175eb-6c81-43c8-97d4-ed36ac824321-cert\") pod \"ingress-canary-zk5kl\" (UID: \"c0d175eb-6c81-43c8-97d4-ed36ac824321\") " pod="openshift-ingress-canary/ingress-canary-zk5kl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.898418 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfhhk\" (UniqueName: \"kubernetes.io/projected/0445a35a-4d10-444b-99ba-b0ced7f82f09-kube-api-access-rfhhk\") pod \"multus-admission-controller-857f4d67dd-lw2ss\" (UID: \"0445a35a-4d10-444b-99ba-b0ced7f82f09\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.929696 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkns5\" (UniqueName: \"kubernetes.io/projected/a41ab691-ce98-4128-a110-6fb02708545e-kube-api-access-kkns5\") pod \"cluster-samples-operator-665b6dd947-4pkp6\" (UID: \"a41ab691-ce98-4128-a110-6fb02708545e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.945368 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ldtm\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-kube-api-access-5ldtm\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.962738 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rg5q9\" (UID: \"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.972141 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:45 crc kubenswrapper[4876]: E0313 07:42:45.972658 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.472636388 +0000 UTC m=+226.143415370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.985779 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d26n\" (UniqueName: \"kubernetes.io/projected/906277a0-614e-442f-873f-4d1d67e4ae5d-kube-api-access-4d26n\") pod \"openshift-controller-manager-operator-756b6f6bc6-zcn4z\" (UID: \"906277a0-614e-442f-873f-4d1d67e4ae5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.988723 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" Mar 13 07:42:45 crc kubenswrapper[4876]: I0313 07:42:45.993861 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.013883 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g8z6\" (UniqueName: \"kubernetes.io/projected/e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f-kube-api-access-2g8z6\") pod \"dns-operator-744455d44c-tpr6j\" (UID: \"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f\") " pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.021989 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.033507 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsk25\" (UniqueName: \"kubernetes.io/projected/ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2-kube-api-access-rsk25\") pod \"package-server-manager-789f6589d5-wbv9l\" (UID: \"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.036528 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpvw6\" (UniqueName: \"kubernetes.io/projected/c59f72f6-172a-4b6e-a367-470966fdb1c1-kube-api-access-hpvw6\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.045799 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.059583 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.063219 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.067338 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnljr\" (UniqueName: \"kubernetes.io/projected/efc8b898-1330-4381-b780-ee41a9514f05-kube-api-access-bnljr\") pod \"collect-profiles-29556450-9j8xm\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.069086 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.073965 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.074801 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.574785386 +0000 UTC m=+226.245564368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.086965 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb28d\" (UniqueName: \"kubernetes.io/projected/f938364d-5a03-47e8-b863-edf8d217770a-kube-api-access-sb28d\") pod \"packageserver-d55dfcdfc-dnkf8\" (UID: \"f938364d-5a03-47e8-b863-edf8d217770a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.088423 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6cnvw"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.094132 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.101411 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.112191 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.113820 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-g65sf"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.117743 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cd4ct" event={"ID":"1fe6fb32-9851-48de-88f6-94701fcf7010","Type":"ContainerStarted","Data":"28de6f41bde64bb52a5ba2f6c6afbe1b29392aa26e0f1231cb73c05397189504"} Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.117785 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cd4ct" event={"ID":"1fe6fb32-9851-48de-88f6-94701fcf7010","Type":"ContainerStarted","Data":"8c12e4f91be7a690b5a0df7b51d7e79a0f80a93d71b86b7a6164d11b755bba45"} Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.118899 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c59f72f6-172a-4b6e-a367-470966fdb1c1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zhz5h\" (UID: \"c59f72f6-172a-4b6e-a367-470966fdb1c1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.123905 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.126214 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" event={"ID":"5e0a197c-3a94-4426-a492-d541a18194a0","Type":"ContainerStarted","Data":"1b4bca46d9763df60de90d0f1ca809cf923cf0b693ec3822a5479b571df0a3a2"} Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.127667 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" event={"ID":"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4","Type":"ContainerStarted","Data":"2ea988a0cbb0462647904994ddfd2d385a2b884e4ae30876c1f152c2dfccfe74"} Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.129506 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" event={"ID":"6b92045d-38c7-4dff-be0a-78449d870bcb","Type":"ContainerStarted","Data":"5949fb0f1aa6fff520cf6b58fce43527d8f22fb89616ab9754b9ee529f10d530"} Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.146115 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lqr7r"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.146913 4876 generic.go:334] "Generic (PLEG): container finished" podID="5461c713-e58f-46e8-91f7-5d6546011e27" containerID="7e558c4e3186174b10f8f552b28e4f1e473ab56c6cc9090959393bc50e943674" exitCode=0 Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.146940 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" event={"ID":"5461c713-e58f-46e8-91f7-5d6546011e27","Type":"ContainerDied","Data":"7e558c4e3186174b10f8f552b28e4f1e473ab56c6cc9090959393bc50e943674"} Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.146958 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" event={"ID":"5461c713-e58f-46e8-91f7-5d6546011e27","Type":"ContainerStarted","Data":"cc636e8704c8c1039589602a55bd37772fd44761db79934b7aa5522efc495aa2"} Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.155002 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw8lv\" (UniqueName: \"kubernetes.io/projected/4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8-kube-api-access-hw8lv\") pod \"service-ca-operator-777779d784-2bszj\" (UID: \"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.164149 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qbl6\" (UniqueName: \"kubernetes.io/projected/c0d175eb-6c81-43c8-97d4-ed36ac824321-kube-api-access-6qbl6\") pod \"ingress-canary-zk5kl\" (UID: \"c0d175eb-6c81-43c8-97d4-ed36ac824321\") " pod="openshift-ingress-canary/ingress-canary-zk5kl" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.177813 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.178555 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zk5kl" Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.179474 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.678441757 +0000 UTC m=+226.349220739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.197045 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxbvz\" (UniqueName: \"kubernetes.io/projected/ed1931f4-3008-4ee4-8295-0c0c99669cf1-kube-api-access-jxbvz\") pod \"dns-default-dfkkx\" (UID: \"ed1931f4-3008-4ee4-8295-0c0c99669cf1\") " pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.200154 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njsxl\" (UniqueName: \"kubernetes.io/projected/38a6f1e3-c474-4f1d-961a-e147adfcbe72-kube-api-access-njsxl\") pod \"machine-config-server-vnfvs\" (UID: \"38a6f1e3-c474-4f1d-961a-e147adfcbe72\") " pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.225075 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54fwt\" (UniqueName: \"kubernetes.io/projected/aa7285c8-4679-495d-afca-6d15a19cfa4c-kube-api-access-54fwt\") pod \"csi-hostpathplugin-m2gxf\" (UID: \"aa7285c8-4679-495d-afca-6d15a19cfa4c\") " pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.279135 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.281485 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.781469849 +0000 UTC m=+226.452248831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.282524 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.283607 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.283639 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.300862 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.310652 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.351582 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.382608 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556462-xxthz"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.383032 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.383499 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.883479863 +0000 UTC m=+226.554258845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.385097 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.398797 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.438623 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.445330 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.452142 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.452616 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nhkg4"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.457846 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xvwj2"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.474049 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.474131 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nkhkk"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.475468 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6k6wd"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.484626 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.490821 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.493512 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vnfvs" Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.493557 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:46.993541181 +0000 UTC m=+226.664320163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.494672 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn"] Mar 13 07:42:46 crc kubenswrapper[4876]: W0313 07:42:46.511057 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod418bec94_3753_4d76_b2a8_24b04f39c820.slice/crio-93909e73d651dc7083069ca32b5ce2c8e6a198e45c349c01536a798d25756586 WatchSource:0}: Error finding container 93909e73d651dc7083069ca32b5ce2c8e6a198e45c349c01536a798d25756586: Status 404 returned error can't find the container with id 93909e73d651dc7083069ca32b5ce2c8e6a198e45c349c01536a798d25756586 Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.515879 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5rhkc"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.525841 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.527632 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g"] Mar 13 07:42:46 crc kubenswrapper[4876]: W0313 07:42:46.571175 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2037c5ee_5855_4520_af01_000ac44338bb.slice/crio-30e3a78c389372d3cafa09944d0e788d1399d408f3374aea8ffb8231f082cd84 WatchSource:0}: Error finding container 30e3a78c389372d3cafa09944d0e788d1399d408f3374aea8ffb8231f082cd84: Status 404 returned error can't find the container with id 30e3a78c389372d3cafa09944d0e788d1399d408f3374aea8ffb8231f082cd84 Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.575540 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.599525 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.599627 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.099605382 +0000 UTC m=+226.770384364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.599776 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.600044 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.100036555 +0000 UTC m=+226.770815527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.665063 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.699168 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.700499 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.700675 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.200583266 +0000 UTC m=+226.871362248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.700874 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.701196 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.201186783 +0000 UTC m=+226.871965765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.785764 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-75kbq"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.790783 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.800501 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6"] Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.802103 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.802417 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.302404223 +0000 UTC m=+226.973183205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: I0313 07:42:46.903911 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:46 crc kubenswrapper[4876]: E0313 07:42:46.905378 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.405357035 +0000 UTC m=+227.076136027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:46 crc kubenswrapper[4876]: W0313 07:42:46.985655 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38a6f1e3_c474_4f1d_961a_e147adfcbe72.slice/crio-dccc2b72e94620f68768da518414ba9cea9ba45539daa907f1d15d8ef508172f WatchSource:0}: Error finding container dccc2b72e94620f68768da518414ba9cea9ba45539daa907f1d15d8ef508172f: Status 404 returned error can't find the container with id dccc2b72e94620f68768da518414ba9cea9ba45539daa907f1d15d8ef508172f Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.006507 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.007235 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.507196553 +0000 UTC m=+227.177975535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.083680 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.084001 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lw2ss"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.084012 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.117475 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.117820 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.617807497 +0000 UTC m=+227.288586479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.160895 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tpr6j"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.174858 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" event={"ID":"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4","Type":"ContainerStarted","Data":"6037594e837a9cd91e7d1250c22248b8e13682f7449755821342a68bb39ca5b3"} Mar 13 07:42:47 crc kubenswrapper[4876]: W0313 07:42:47.176830 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb82a006d_154c_4e07_b683_c65ca4863459.slice/crio-be3e5a7ea54fb7ba95ba0f9684533f57c6f6f16c91195c64c1fa1fe1430d5976 WatchSource:0}: Error finding container be3e5a7ea54fb7ba95ba0f9684533f57c6f6f16c91195c64c1fa1fe1430d5976: Status 404 returned error can't find the container with id be3e5a7ea54fb7ba95ba0f9684533f57c6f6f16c91195c64c1fa1fe1430d5976 Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.185402 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" event={"ID":"29aded1a-dd1a-4ce1-b441-b3b897e4a06d","Type":"ContainerStarted","Data":"4b589f3fa4cc3df5da25f98ddfa4361416084d29ffa67af3fe06c4f10d39f1ea"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.189260 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g65sf" event={"ID":"62fdbab2-2be8-42cc-a936-a87ffd085a0e","Type":"ContainerStarted","Data":"02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.189319 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g65sf" event={"ID":"62fdbab2-2be8-42cc-a936-a87ffd085a0e","Type":"ContainerStarted","Data":"7f68a1abd786e08ed55e9e6f82c3c8fb64fbb9e19e00d648e05cf0ab70d47a54"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.200846 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" event={"ID":"418bec94-3753-4d76-b2a8-24b04f39c820","Type":"ContainerStarted","Data":"dbe168b993e7208f1f475e8dae634b31de4acc14e721ae31cdaeb38994cd92fa"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.200892 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" event={"ID":"418bec94-3753-4d76-b2a8-24b04f39c820","Type":"ContainerStarted","Data":"93909e73d651dc7083069ca32b5ce2c8e6a198e45c349c01536a798d25756586"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.217940 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.219293 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" event={"ID":"945666f1-2703-4fa7-9063-fd49b7189341","Type":"ContainerStarted","Data":"e2809c40ce0ec4902076cf7982e9111be8d198da1dd569c46fe8533f9b9c72ed"} Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.219379 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.719363297 +0000 UTC m=+227.390142279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.220055 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.221680 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.721670844 +0000 UTC m=+227.392449826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.223935 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.233732 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" event={"ID":"6f60bac9-dca3-4eb8-b163-8146b084e74d","Type":"ContainerStarted","Data":"7506b4049f71059055960bf7b70a33e7cdcc92d98c436406f24486c71cbab448"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.234448 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.238621 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.240721 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.241661 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.241685 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" event={"ID":"300d0a76-b532-4124-af3d-f89cf7a51d3f","Type":"ContainerStarted","Data":"e85adc5bed20d4124bc4963d768f395c1a9c9739930b4c2b454f7094cca44e3b"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.247454 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" event={"ID":"6b92045d-38c7-4dff-be0a-78449d870bcb","Type":"ContainerStarted","Data":"6902aef66af563dcddb390d4ca47a333014cfe448070e37021e928c7d4e9e200"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.248889 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.249902 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" event={"ID":"90d11aee-636b-4d59-bb06-017a3adafe81","Type":"ContainerStarted","Data":"660bf4923b7442e3c0be21d23a5e1705dd4e3390678fa5f31bda121a8726cf35"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.251160 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vnfvs" event={"ID":"38a6f1e3-c474-4f1d-961a-e147adfcbe72","Type":"ContainerStarted","Data":"dccc2b72e94620f68768da518414ba9cea9ba45539daa907f1d15d8ef508172f"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.251906 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" event={"ID":"733daf88-a275-4be5-ace0-0d12edaf1b87","Type":"ContainerStarted","Data":"993be9879b4702233c65d1f989f7839f0bc43c0b727a703d130443b128ab89e7"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.252396 4876 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-xvfbt container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.252434 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" podUID="6b92045d-38c7-4dff-be0a-78449d870bcb" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.261649 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" event={"ID":"15c836ec-fd1e-47ee-bfc7-90c419082e8b","Type":"ContainerStarted","Data":"f49eb383c868dc2fbc5fcae57aa8cffb7736e8a145cdd3f884fcfafa8ed39696"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.263716 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" event={"ID":"906277a0-614e-442f-873f-4d1d67e4ae5d","Type":"ContainerStarted","Data":"f4d6e39342613bfa306164c7742cd7ab36e367b6c5bb2d6ef3fa79897e776518"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.271347 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556462-xxthz" event={"ID":"01e125dd-87e2-4362-842e-849288698994","Type":"ContainerStarted","Data":"005a74e6f6269bd9bf747ff7c6885857c52d9bda3e2efb70a162e8e5b4e88f6c"} Mar 13 07:42:47 crc kubenswrapper[4876]: W0313 07:42:47.271985 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0a8ecd9_f9f1_4114_a5d1_8fb53262ca3f.slice/crio-a11e16d893e1c838b793fdd790a297df862ccf9f1e9820bfcbda5ac14bbd6f95 WatchSource:0}: Error finding container a11e16d893e1c838b793fdd790a297df862ccf9f1e9820bfcbda5ac14bbd6f95: Status 404 returned error can't find the container with id a11e16d893e1c838b793fdd790a297df862ccf9f1e9820bfcbda5ac14bbd6f95 Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.275068 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m2gxf"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.289387 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:47 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:47 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:47 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.289448 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.290198 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" event={"ID":"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2","Type":"ContainerStarted","Data":"220199b0d93904fe88f39ad029661bfb85b3c17beaf3cafd29e88a02638a2d9d"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.295974 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" event={"ID":"f089ab79-2e29-4074-a8dc-21b0af79d3f7","Type":"ContainerStarted","Data":"052a6b45f498fd6d280d3e717ed76cb57f975004a439506fd4c7b4ab3f00d161"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.297900 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" event={"ID":"2037c5ee-5855-4520-af01-000ac44338bb","Type":"ContainerStarted","Data":"30e3a78c389372d3cafa09944d0e788d1399d408f3374aea8ffb8231f082cd84"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.312621 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" event={"ID":"f8b52e5e-95cb-4d81-b3c2-d014955b31fd","Type":"ContainerStarted","Data":"fd4d3b5581750eb5c9411366d6355ea96e64f230ff0f8c6da5ed255d457b8f00"} Mar 13 07:42:47 crc kubenswrapper[4876]: W0313 07:42:47.318380 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6eb3c1a_abbc_4b82_950e_e3cde8a5da8a.slice/crio-421bf27305fc646e14505de75a2b3781fc0feb0f54003e05129e249ae985af02 WatchSource:0}: Error finding container 421bf27305fc646e14505de75a2b3781fc0feb0f54003e05129e249ae985af02: Status 404 returned error can't find the container with id 421bf27305fc646e14505de75a2b3781fc0feb0f54003e05129e249ae985af02 Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.321016 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.322255 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.822213705 +0000 UTC m=+227.492992687 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.325644 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zk5kl"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.327712 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dfkkx"] Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.336579 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6cnvw" event={"ID":"ee3187e1-6603-4cd1-a134-6437fae1ca4a","Type":"ContainerStarted","Data":"ddea5c6d3e631d9d574fe45141db5454e8f2d38a73b22d3248b2de0c95e65212"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.336630 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6cnvw" event={"ID":"ee3187e1-6603-4cd1-a134-6437fae1ca4a","Type":"ContainerStarted","Data":"52b9c3dc87c0fbe66f04ceebf927f185fb29cd738cd0260e50b9eb248ba56284"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.337254 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.338825 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" event={"ID":"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea","Type":"ContainerStarted","Data":"1fe6b1cccf817fbb0459285fa8619378bcaf32140b1359069a2cc68242e4dd9b"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.340428 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" event={"ID":"918fce5c-2c8a-46a8-9c16-3911fb36faa3","Type":"ContainerStarted","Data":"f70af5fa74372148fadf041fcc62c0fc0c4882dce1d40a4fa176596c30615db1"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.341609 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" event={"ID":"0a8ea6bd-3402-40ee-b509-51c54f8af982","Type":"ContainerStarted","Data":"831730dd37bef3eca54bf2839a8255773452b38bb9e428d3fff224f0df0aeb05"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.343467 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" event={"ID":"caab36f1-979c-459d-b742-941e13dc765e","Type":"ContainerStarted","Data":"44b01e5579ce59db544b51b114195816687bda9064b8383e2e0e0d386019f73a"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.345291 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" event={"ID":"7984a28c-6710-47b1-9c43-7e2c39dda944","Type":"ContainerStarted","Data":"85d00e11e739f386b75fc8c307bb5473c7532f44e202e33b729cdca18b8b9ae5"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.351933 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.351999 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.365576 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" event={"ID":"99f7833f-00a4-45fe-80ba-6c22404c49e4","Type":"ContainerStarted","Data":"a394170f18f50ef6ec87aaf42054d5ef3231f6f4d0bec344aca619286d2b8f3a"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.365624 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" event={"ID":"99f7833f-00a4-45fe-80ba-6c22404c49e4","Type":"ContainerStarted","Data":"65242d0a23f5b6e2366d8a320bed16ea6fac31c22f74f35797abd740de766800"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.380442 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" event={"ID":"8c6d92f4-c514-48e4-b1da-61546f07d171","Type":"ContainerStarted","Data":"b6890891631d9c33c93d014b2e50a8351387e6ea745a9b7beae2104bdf6805ad"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.383745 4876 generic.go:334] "Generic (PLEG): container finished" podID="5e0a197c-3a94-4426-a492-d541a18194a0" containerID="20a3bd2c0332ecc4cb9b6e3b6ce2cf6d1add64079fcbe52c19f93dc80b0c15e5" exitCode=0 Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.383819 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" event={"ID":"5e0a197c-3a94-4426-a492-d541a18194a0","Type":"ContainerDied","Data":"20a3bd2c0332ecc4cb9b6e3b6ce2cf6d1add64079fcbe52c19f93dc80b0c15e5"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.393368 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" event={"ID":"a41ab691-ce98-4128-a110-6fb02708545e","Type":"ContainerStarted","Data":"35e1268d5069769df9ecbbb83aa158e8b1f978a7c27a94fccf7ccb280c1d12e5"} Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.411536 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2bszj"] Mar 13 07:42:47 crc kubenswrapper[4876]: W0313 07:42:47.417014 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded1931f4_3008_4ee4_8295_0c0c99669cf1.slice/crio-3f02b095cd1dff34537e2cb81edce559ee53569b73007a575e32aeaa9cba2774 WatchSource:0}: Error finding container 3f02b095cd1dff34537e2cb81edce559ee53569b73007a575e32aeaa9cba2774: Status 404 returned error can't find the container with id 3f02b095cd1dff34537e2cb81edce559ee53569b73007a575e32aeaa9cba2774 Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.425870 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.429674 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:47.929656066 +0000 UTC m=+227.600435048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: W0313 07:42:47.519508 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d355f08_2e01_4102_8cd4_b6f3b8e8b3d8.slice/crio-5fa3a0847085a4ff6181d3b1524f905533e5c79ee3b50eb1a69aebda93cadc9b WatchSource:0}: Error finding container 5fa3a0847085a4ff6181d3b1524f905533e5c79ee3b50eb1a69aebda93cadc9b: Status 404 returned error can't find the container with id 5fa3a0847085a4ff6181d3b1524f905533e5c79ee3b50eb1a69aebda93cadc9b Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.541676 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.541844 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.041820204 +0000 UTC m=+227.712599186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.541972 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.542686 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.042669379 +0000 UTC m=+227.713448431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.639994 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-cd4ct" podStartSLOduration=167.639972197 podStartE2EDuration="2m47.639972197s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:47.59706312 +0000 UTC m=+227.267842102" watchObservedRunningTime="2026-03-13 07:42:47.639972197 +0000 UTC m=+227.310751179" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.642958 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.643281 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.143265812 +0000 UTC m=+227.814044794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.747293 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.747833 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.24781565 +0000 UTC m=+227.918594702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.831670 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6cnvw" podStartSLOduration=168.831651534 podStartE2EDuration="2m48.831651534s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:47.79636716 +0000 UTC m=+227.467146152" watchObservedRunningTime="2026-03-13 07:42:47.831651534 +0000 UTC m=+227.502430516" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.834050 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbjjp" podStartSLOduration=168.834044084 podStartE2EDuration="2m48.834044084s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:47.832443908 +0000 UTC m=+227.503222890" watchObservedRunningTime="2026-03-13 07:42:47.834044084 +0000 UTC m=+227.504823056" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.848050 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.848441 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.348420242 +0000 UTC m=+228.019199224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.919127 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-g65sf" podStartSLOduration=168.919109526 podStartE2EDuration="2m48.919109526s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:47.917990793 +0000 UTC m=+227.588769775" watchObservedRunningTime="2026-03-13 07:42:47.919109526 +0000 UTC m=+227.589888508" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.919710 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" podStartSLOduration=168.919705613 podStartE2EDuration="2m48.919705613s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:47.881184404 +0000 UTC m=+227.551963386" watchObservedRunningTime="2026-03-13 07:42:47.919705613 +0000 UTC m=+227.590484595" Mar 13 07:42:47 crc kubenswrapper[4876]: I0313 07:42:47.953190 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:47 crc kubenswrapper[4876]: E0313 07:42:47.953584 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.453570127 +0000 UTC m=+228.124349109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.055578 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.056357 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.556337563 +0000 UTC m=+228.227116545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.159870 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.160412 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.660392215 +0000 UTC m=+228.331171197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.262324 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.262872 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.762855482 +0000 UTC m=+228.433634464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.295383 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:48 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:48 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:48 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.295464 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.364065 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.364495 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.864474604 +0000 UTC m=+228.535253586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.432642 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" event={"ID":"906277a0-614e-442f-873f-4d1d67e4ae5d","Type":"ContainerStarted","Data":"d5d39c2d437679b2f0ce64bd17dd9ebbd9ae19d302f342b2f3dcbb6d00e0e045"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.462076 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" event={"ID":"945666f1-2703-4fa7-9063-fd49b7189341","Type":"ContainerStarted","Data":"9be711b58b5894c90ca45ab173bdf71ff3eb5405503c8a6f03ee94f552507841"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.463157 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.464500 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.464832 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:48.96481982 +0000 UTC m=+228.635598802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.464894 4876 patch_prober.go:28] interesting pod/console-operator-58897d9998-5rhkc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.464918 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" podUID="945666f1-2703-4fa7-9063-fd49b7189341" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.466214 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" event={"ID":"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f","Type":"ContainerStarted","Data":"a11e16d893e1c838b793fdd790a297df862ccf9f1e9820bfcbda5ac14bbd6f95"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.474344 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" event={"ID":"c59f72f6-172a-4b6e-a367-470966fdb1c1","Type":"ContainerStarted","Data":"38ebd2a285b2c17e7e0d82c40f4754abcbb2947354b0acd4101425ab77731028"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.475812 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" event={"ID":"418bec94-3753-4d76-b2a8-24b04f39c820","Type":"ContainerStarted","Data":"ac44ae0d10075fc0e01da7043c4d334d961acc7fd2877f58f91790e5e293fcd3"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.477682 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" event={"ID":"90d11aee-636b-4d59-bb06-017a3adafe81","Type":"ContainerStarted","Data":"a0b226c1f63a7d9e0d6cbc11bbab9c2a993c32bb9cc0ea82d62881d3d498741c"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.478224 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.479057 4876 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-vwq9g container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.479097 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" podUID="90d11aee-636b-4d59-bb06-017a3adafe81" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.488605 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" event={"ID":"6f60bac9-dca3-4eb8-b163-8146b084e74d","Type":"ContainerStarted","Data":"781b07e2c922a221c3e0dcc46c95d75b2d2656ed0198b8cd14c576630694f97d"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.510558 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zcn4z" podStartSLOduration=169.510538247 podStartE2EDuration="2m49.510538247s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:48.505973785 +0000 UTC m=+228.176752787" watchObservedRunningTime="2026-03-13 07:42:48.510538247 +0000 UTC m=+228.181317229" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.526391 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" event={"ID":"15c836ec-fd1e-47ee-bfc7-90c419082e8b","Type":"ContainerStarted","Data":"6947c47ffa04541ad12b9dd3614ef1a703f85fb8b0f4cc97abe233bf22e7c128"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.552190 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" event={"ID":"0a8ea6bd-3402-40ee-b509-51c54f8af982","Type":"ContainerStarted","Data":"316dd098f840c7ec7d6511c42ef59cb9b24ef3b7e8f48177c3e236291be52ff4"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.565837 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.566769 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.066758231 +0000 UTC m=+228.737537213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.569208 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" event={"ID":"efc8b898-1330-4381-b780-ee41a9514f05","Type":"ContainerStarted","Data":"b259609c4cbb383974b53f6f1a6384734c1f9ebf217a9b1f742f3b6812c86700"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.603897 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" event={"ID":"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2","Type":"ContainerStarted","Data":"cc6ed110258c6762ee845c2f13301516f9c7b773ae67bc9da6bd7e59046c69ce"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.605575 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.610261 4876 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nkhkk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.610327 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" podUID="ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.670520 4876 generic.go:334] "Generic (PLEG): container finished" podID="caab36f1-979c-459d-b742-941e13dc765e" containerID="6f6bfd5b38316a430550122a8735e1fc74f37ec97627fd32b2d404769e60edf4" exitCode=0 Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.670616 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" event={"ID":"caab36f1-979c-459d-b742-941e13dc765e","Type":"ContainerDied","Data":"6f6bfd5b38316a430550122a8735e1fc74f37ec97627fd32b2d404769e60edf4"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.672137 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.674276 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.174228573 +0000 UTC m=+228.845007555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.722119 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" podStartSLOduration=168.722092553 podStartE2EDuration="2m48.722092553s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:48.670212406 +0000 UTC m=+228.340991388" watchObservedRunningTime="2026-03-13 07:42:48.722092553 +0000 UTC m=+228.392871535" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.778285 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.778554 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" event={"ID":"f089ab79-2e29-4074-a8dc-21b0af79d3f7","Type":"ContainerStarted","Data":"647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.779697 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.779939 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.279923414 +0000 UTC m=+228.950702396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.795695 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-xvwj2" podStartSLOduration=168.795667761 podStartE2EDuration="2m48.795667761s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:48.791905332 +0000 UTC m=+228.462684334" watchObservedRunningTime="2026-03-13 07:42:48.795667761 +0000 UTC m=+228.466446743" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.797214 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" podStartSLOduration=169.797207836 podStartE2EDuration="2m49.797207836s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:48.723803124 +0000 UTC m=+228.394582106" watchObservedRunningTime="2026-03-13 07:42:48.797207836 +0000 UTC m=+228.467986828" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.813284 4876 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-75kbq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.813348 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" podUID="f089ab79-2e29-4074-a8dc-21b0af79d3f7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.822191 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dfkkx" event={"ID":"ed1931f4-3008-4ee4-8295-0c0c99669cf1","Type":"ContainerStarted","Data":"3f02b095cd1dff34537e2cb81edce559ee53569b73007a575e32aeaa9cba2774"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.860689 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" event={"ID":"b82a006d-154c-4e07-b683-c65ca4863459","Type":"ContainerStarted","Data":"be3e5a7ea54fb7ba95ba0f9684533f57c6f6f16c91195c64c1fa1fe1430d5976"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.886264 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.888757 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.388732694 +0000 UTC m=+229.059511676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.888928 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" event={"ID":"a41ab691-ce98-4128-a110-6fb02708545e","Type":"ContainerStarted","Data":"49747d3ec1260103e0f3c1ebb8537b8d8685c2de3428d53c7009fb000a5cadb3"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.893045 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" event={"ID":"5461c713-e58f-46e8-91f7-5d6546011e27","Type":"ContainerStarted","Data":"91556f57165b51427919f75c5c059fa12017b962a508cbf91cb79a4a2be9a17e"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.899976 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" podStartSLOduration=169.899960441 podStartE2EDuration="2m49.899960441s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:48.897904632 +0000 UTC m=+228.568683614" watchObservedRunningTime="2026-03-13 07:42:48.899960441 +0000 UTC m=+228.570739423" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.900398 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" podStartSLOduration=169.900392684 podStartE2EDuration="2m49.900392684s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:48.860974209 +0000 UTC m=+228.531753181" watchObservedRunningTime="2026-03-13 07:42:48.900392684 +0000 UTC m=+228.571171666" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.901036 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" event={"ID":"aa7285c8-4679-495d-afca-6d15a19cfa4c","Type":"ContainerStarted","Data":"2d93b8e7c7923faad162adb32fc2966ea1a3cfa776ec34c93e04730227c38701"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.914185 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" event={"ID":"f938364d-5a03-47e8-b863-edf8d217770a","Type":"ContainerStarted","Data":"39cbdda8c7c51d5c0568bbc8e502cdb56599109950687b0848762f6efd38ccdd"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.914267 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" event={"ID":"f938364d-5a03-47e8-b863-edf8d217770a","Type":"ContainerStarted","Data":"6fa978bf6ffbc029f325870d97e3990e5cae08fee47e4388f479fa9f27ed8fd7"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.918682 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.923390 4876 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dnkf8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.923451 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" podUID="f938364d-5a03-47e8-b863-edf8d217770a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.929850 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dlvtf" podStartSLOduration=168.929813229 podStartE2EDuration="2m48.929813229s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:48.918931162 +0000 UTC m=+228.589710144" watchObservedRunningTime="2026-03-13 07:42:48.929813229 +0000 UTC m=+228.600592231" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.975540 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vnfvs" event={"ID":"38a6f1e3-c474-4f1d-961a-e147adfcbe72","Type":"ContainerStarted","Data":"946286bd8eb4477ca395c8951d3c3515139445fcdf1448143dcb47eca90acc97"} Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.982637 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" podStartSLOduration=168.982614572 podStartE2EDuration="2m48.982614572s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:48.981640485 +0000 UTC m=+228.652419467" watchObservedRunningTime="2026-03-13 07:42:48.982614572 +0000 UTC m=+228.653393554" Mar 13 07:42:48 crc kubenswrapper[4876]: I0313 07:42:48.988098 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:48 crc kubenswrapper[4876]: E0313 07:42:48.989074 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.48905881 +0000 UTC m=+229.159837792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:48.995106 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" event={"ID":"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a","Type":"ContainerStarted","Data":"9eb955c8b407710c278c4780cb0c3343cce09e6daa3ac6cecd2b2102bd4eb8f9"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:48.995176 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" event={"ID":"b6eb3c1a-abbc-4b82-950e-e3cde8a5da8a","Type":"ContainerStarted","Data":"421bf27305fc646e14505de75a2b3781fc0feb0f54003e05129e249ae985af02"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.006019 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" event={"ID":"f8b52e5e-95cb-4d81-b3c2-d014955b31fd","Type":"ContainerStarted","Data":"c0dd9b9df06a80699c8834997b70e44c27af2e486bb85af9a12062f486f3abb7"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.006095 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" podStartSLOduration=169.006079974 podStartE2EDuration="2m49.006079974s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.004128978 +0000 UTC m=+228.674907960" watchObservedRunningTime="2026-03-13 07:42:49.006079974 +0000 UTC m=+228.676858956" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.006568 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.008537 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zk5kl" event={"ID":"c0d175eb-6c81-43c8-97d4-ed36ac824321","Type":"ContainerStarted","Data":"48d58dc14ff59a6fa3a4234de1b322292bc25f6bf6bf31c5a922506174f310aa"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.025272 4876 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-jdqxn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.025338 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" podUID="f8b52e5e-95cb-4d81-b3c2-d014955b31fd" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.044405 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-vnfvs" podStartSLOduration=6.044375567 podStartE2EDuration="6.044375567s" podCreationTimestamp="2026-03-13 07:42:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.034571212 +0000 UTC m=+228.705350204" watchObservedRunningTime="2026-03-13 07:42:49.044375567 +0000 UTC m=+228.715154549" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.088596 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" event={"ID":"8c6d92f4-c514-48e4-b1da-61546f07d171","Type":"ContainerStarted","Data":"1c195b6fa4144f07b4e696b40cec201b35aa726d34e069dfd1293f96ed319d2b"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.089053 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.091706 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.591688191 +0000 UTC m=+229.262467173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.093543 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" podStartSLOduration=169.093524795 podStartE2EDuration="2m49.093524795s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.092775863 +0000 UTC m=+228.763554835" watchObservedRunningTime="2026-03-13 07:42:49.093524795 +0000 UTC m=+228.764303777" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.094973 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rg5q9" podStartSLOduration=169.094966587 podStartE2EDuration="2m49.094966587s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.059154966 +0000 UTC m=+228.729933948" watchObservedRunningTime="2026-03-13 07:42:49.094966587 +0000 UTC m=+228.765745569" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.097145 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" event={"ID":"34f1aacf-f7ff-4f62-99e6-1826c7f7c5f4","Type":"ContainerStarted","Data":"6db395140bc23bd63e6b786909399daac33c0a120c80ea7eff17884bfdc52a6b"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.123899 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" event={"ID":"918fce5c-2c8a-46a8-9c16-3911fb36faa3","Type":"ContainerStarted","Data":"a3a554096ff27a4304c7e73de300a3aade799e39229774f2e8713f45c521b474"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.194088 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.195371 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hln48" podStartSLOduration=169.195355733 podStartE2EDuration="2m49.195355733s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.134830355 +0000 UTC m=+228.805609337" watchObservedRunningTime="2026-03-13 07:42:49.195355733 +0000 UTC m=+228.866134715" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.196982 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-lqr7r" podStartSLOduration=170.196977061 podStartE2EDuration="2m50.196977061s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.194782197 +0000 UTC m=+228.865561179" watchObservedRunningTime="2026-03-13 07:42:49.196977061 +0000 UTC m=+228.867756043" Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.197490 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.697472325 +0000 UTC m=+229.368251387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.272938 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" event={"ID":"29aded1a-dd1a-4ce1-b441-b3b897e4a06d","Type":"ContainerStarted","Data":"a1bc04e55e62e4790f61fe7a1190bf8e7a2b1520463b1abc4873c045aa617a64"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.295872 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.296466 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.796449301 +0000 UTC m=+229.467228293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.300196 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:49 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:49 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:49 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.300270 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.321586 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" event={"ID":"2037c5ee-5855-4520-af01-000ac44338bb","Type":"ContainerStarted","Data":"891d223b64475a664b84363f9f5779fa716481a14b44ac459ebcebab5bcbdd82"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.326089 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mb8k7" podStartSLOduration=170.326067481 podStartE2EDuration="2m50.326067481s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.316820552 +0000 UTC m=+228.987599534" watchObservedRunningTime="2026-03-13 07:42:49.326067481 +0000 UTC m=+228.996846483" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.326215 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dnr6c" podStartSLOduration=170.326211555 podStartE2EDuration="2m50.326211555s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.235727726 +0000 UTC m=+228.906506708" watchObservedRunningTime="2026-03-13 07:42:49.326211555 +0000 UTC m=+228.996990537" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.346472 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" event={"ID":"300d0a76-b532-4124-af3d-f89cf7a51d3f","Type":"ContainerStarted","Data":"aaf2181ac87531546d71335703d83d8c0bdd8cc711be50775a867db6da83336f"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.348016 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.352010 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" event={"ID":"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2","Type":"ContainerStarted","Data":"7677e12b231f4b35a99e74279fd368ecf11dd7870640e3166bbe8d8339c7b7f8"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.356183 4876 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-z8697 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.356269 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" podUID="300d0a76-b532-4124-af3d-f89cf7a51d3f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.358554 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" event={"ID":"0445a35a-4d10-444b-99ba-b0ced7f82f09","Type":"ContainerStarted","Data":"496e804334afcc36af9a11b08b6ed8371915ad45c9a61b348f71a85314658b20"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.387119 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-6k6wd" podStartSLOduration=169.387092693 podStartE2EDuration="2m49.387092693s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.378690889 +0000 UTC m=+229.049469891" watchObservedRunningTime="2026-03-13 07:42:49.387092693 +0000 UTC m=+229.057871675" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.400039 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.401408 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:49.901388339 +0000 UTC m=+229.572167321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.426118 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" podStartSLOduration=169.426098577 podStartE2EDuration="2m49.426098577s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.425471369 +0000 UTC m=+229.096250351" watchObservedRunningTime="2026-03-13 07:42:49.426098577 +0000 UTC m=+229.096877549" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.435748 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" event={"ID":"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8","Type":"ContainerStarted","Data":"5fa3a0847085a4ff6181d3b1524f905533e5c79ee3b50eb1a69aebda93cadc9b"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.448584 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" event={"ID":"733daf88-a275-4be5-ace0-0d12edaf1b87","Type":"ContainerStarted","Data":"2af9d43fad3953061c5ca135a9190bf1babb8eeef3937b6d6368f4b1441ddec2"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.462820 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" event={"ID":"b88bd1db-0ac9-41d9-a862-2bc0b3ae88ea","Type":"ContainerStarted","Data":"fe62ea0967c34abb58cc8ba3f627969ed794ed94b5c1e09b99d6e55d1ccbb9f3"} Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.463832 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.463889 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.478744 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.498129 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" podStartSLOduration=169.498108299 podStartE2EDuration="2m49.498108299s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.482034722 +0000 UTC m=+229.152813724" watchObservedRunningTime="2026-03-13 07:42:49.498108299 +0000 UTC m=+229.168887281" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.535156 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.541357 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.041325425 +0000 UTC m=+229.712104407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.541519 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.546007 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.04599151 +0000 UTC m=+229.716770492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.588884 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pnjqr" podStartSLOduration=169.588858935 podStartE2EDuration="2m49.588858935s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.580387119 +0000 UTC m=+229.251166111" watchObservedRunningTime="2026-03-13 07:42:49.588858935 +0000 UTC m=+229.259637917" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.636816 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hlwjw" podStartSLOduration=169.636800488 podStartE2EDuration="2m49.636800488s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:49.636338734 +0000 UTC m=+229.307117716" watchObservedRunningTime="2026-03-13 07:42:49.636800488 +0000 UTC m=+229.307579470" Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.653118 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.654443 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.15442801 +0000 UTC m=+229.825206992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.755394 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.755751 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.255736484 +0000 UTC m=+229.926515466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.857180 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.857491 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.357463049 +0000 UTC m=+230.028242031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.857594 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.857964 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.357953713 +0000 UTC m=+230.028732705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.959024 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.959166 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.459126672 +0000 UTC m=+230.129905654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:49 crc kubenswrapper[4876]: I0313 07:42:49.959677 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:49 crc kubenswrapper[4876]: E0313 07:42:49.959937 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.459929845 +0000 UTC m=+230.130708817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.062053 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.062476 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.562457865 +0000 UTC m=+230.233236847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.164871 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.165281 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.665265801 +0000 UTC m=+230.336044783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.265570 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.265729 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.765703659 +0000 UTC m=+230.436482641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.265815 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.266151 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.766144182 +0000 UTC m=+230.436923164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.289315 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:50 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:50 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:50 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.289661 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.367380 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.367824 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.867808225 +0000 UTC m=+230.538587207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.468597 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.468916 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:50.968902002 +0000 UTC m=+230.639680984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.491692 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" event={"ID":"7984a28c-6710-47b1-9c43-7e2c39dda944","Type":"ContainerStarted","Data":"b9e16bb19de89ec354be8436b824fb12125580307ae2af89436cde6e395dc680"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.491744 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" event={"ID":"7984a28c-6710-47b1-9c43-7e2c39dda944","Type":"ContainerStarted","Data":"c1d7be156201164bbf4301c6275128601b3bce30a1aea0948428f6bf91bdcc56"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.493578 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" event={"ID":"b82a006d-154c-4e07-b683-c65ca4863459","Type":"ContainerStarted","Data":"a502591803ade5b8f8c0ff451008eb78650524a1228109b0c25a25be7c9e3765"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.493600 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" event={"ID":"b82a006d-154c-4e07-b683-c65ca4863459","Type":"ContainerStarted","Data":"849a82cfd7a0a61021cc81ad8e709c6857771405069ed0393f26056a99a9885e"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.495305 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" event={"ID":"a41ab691-ce98-4128-a110-6fb02708545e","Type":"ContainerStarted","Data":"020852d5eea21552b762a289e7d6e0daa5e174bdd4ec3c23ebfc76ed6524abe6"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.502707 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" event={"ID":"6f60bac9-dca3-4eb8-b163-8146b084e74d","Type":"ContainerStarted","Data":"45208989f8098a699b8017ba234937193af830059b0eb54649159eaf9d04d604"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.505175 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" event={"ID":"caab36f1-979c-459d-b742-941e13dc765e","Type":"ContainerStarted","Data":"948210ade28494fbec69c4f6f67603a4e032cbd41ed5e82eb2e6f297192776f6"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.505587 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.508888 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" event={"ID":"aa7285c8-4679-495d-afca-6d15a19cfa4c","Type":"ContainerStarted","Data":"04723d59ab27077a20bce3b0895b8d2bea4c5a4e73ab5145694ab87fbc1e8e3b"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.510226 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" event={"ID":"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f","Type":"ContainerStarted","Data":"891ea6fcfa93c4bc53a7924e50b7d57671af7a3a100fa6fdbb101558edd2fea6"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.510269 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" event={"ID":"e0a8ecd9-f9f1-4114-a5d1-8fb53262ca3f","Type":"ContainerStarted","Data":"da3d4cdc0a69e7cd07cf160e9c9eae983badd09b48803253de21a642dc9ad894"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.518937 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2bszj" event={"ID":"4d355f08-2e01-4102-8cd4-b6f3b8e8b3d8","Type":"ContainerStarted","Data":"4b87c69df72dd8f19acab6e6d8012bb6b13b08878292364ae5894c9cb4e6a670"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.523982 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zk5kl" event={"ID":"c0d175eb-6c81-43c8-97d4-ed36ac824321","Type":"ContainerStarted","Data":"c1d14456b518fcdc52eb026c7090e67ad656177eef91efd2263c45130b21d80e"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.526698 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dfkkx" event={"ID":"ed1931f4-3008-4ee4-8295-0c0c99669cf1","Type":"ContainerStarted","Data":"a8b8d641dd77e662398f3f3d210d3f3d6eca06c7210df4aa2db005eede94fe23"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.526723 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dfkkx" event={"ID":"ed1931f4-3008-4ee4-8295-0c0c99669cf1","Type":"ContainerStarted","Data":"1e65cbbfd18529593bbaed9f7180ca181b4ed6f57e8968da8efa7f8cb7257af6"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.527173 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.530566 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" event={"ID":"c59f72f6-172a-4b6e-a367-470966fdb1c1","Type":"ContainerStarted","Data":"30be29e040dea24f52a03791bb1e4f815138afba2439d39ef9603350fec083a7"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.530594 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" event={"ID":"c59f72f6-172a-4b6e-a367-470966fdb1c1","Type":"ContainerStarted","Data":"b2ae391746e7a7d35aa9c4100df6d1f0803180a68b61ac000e94fd8363879e8e"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.534925 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xkfqg" podStartSLOduration=170.53490767 podStartE2EDuration="2m50.53490767s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.516785963 +0000 UTC m=+230.187564955" watchObservedRunningTime="2026-03-13 07:42:50.53490767 +0000 UTC m=+230.205686652" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.540787 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" event={"ID":"0445a35a-4d10-444b-99ba-b0ced7f82f09","Type":"ContainerStarted","Data":"4fd6f932dd27b87108bc547930a8eac9e2f00617c837488825ee51134bff50a2"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.560608 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-tpr6j" podStartSLOduration=171.560592765 podStartE2EDuration="2m51.560592765s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.536393473 +0000 UTC m=+230.207172455" watchObservedRunningTime="2026-03-13 07:42:50.560592765 +0000 UTC m=+230.231371747" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.566026 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" event={"ID":"5461c713-e58f-46e8-91f7-5d6546011e27","Type":"ContainerStarted","Data":"d1d63a4800ffd96a786db9c5a396ffc4440b4df4bcbcb5da9589af97a8d0476e"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.568303 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" event={"ID":"5e0a197c-3a94-4426-a492-d541a18194a0","Type":"ContainerStarted","Data":"89071092cbdc85b88fa4eeb6e44979a5e0abaca47efd703eba97da5ec938613f"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.569163 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.569423 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.069410642 +0000 UTC m=+230.740189624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.570197 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" event={"ID":"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2","Type":"ContainerStarted","Data":"41e9e745abb2fc05cc7b1f2546f50268bd1fef1011f497dcf56f218ef1e2deb6"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.570215 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" event={"ID":"ed56e4be-9c0d-4316-afaa-6d5b99ec3bb2","Type":"ContainerStarted","Data":"64559887f2785ae2a875d9b8a993c50638eab7afd662167c8ecb7df7623b1c8b"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.570578 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.590264 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" event={"ID":"efc8b898-1330-4381-b780-ee41a9514f05","Type":"ContainerStarted","Data":"181d010f2fcc26b78278f10c1875f24e976a91b573cefd134991bf5558ea2faa"} Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.592152 4876 patch_prober.go:28] interesting pod/console-operator-58897d9998-5rhkc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.592202 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" podUID="945666f1-2703-4fa7-9063-fd49b7189341" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.592223 4876 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-75kbq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.592284 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" podUID="f089ab79-2e29-4074-a8dc-21b0af79d3f7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.592762 4876 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nkhkk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.592816 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" podUID="ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.596606 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.596646 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.596939 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zktvd" podStartSLOduration=170.596920851 podStartE2EDuration="2m50.596920851s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.561996996 +0000 UTC m=+230.232775988" watchObservedRunningTime="2026-03-13 07:42:50.596920851 +0000 UTC m=+230.267699833" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.597878 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" podStartSLOduration=171.597871379 podStartE2EDuration="2m51.597871379s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.596454768 +0000 UTC m=+230.267233750" watchObservedRunningTime="2026-03-13 07:42:50.597871379 +0000 UTC m=+230.268650361" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.624102 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4pkp6" podStartSLOduration=171.624089451 podStartE2EDuration="2m51.624089451s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.623637348 +0000 UTC m=+230.294416330" watchObservedRunningTime="2026-03-13 07:42:50.624089451 +0000 UTC m=+230.294868433" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.643388 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z8697" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.645869 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn5d4" podStartSLOduration=170.645856672 podStartE2EDuration="2m50.645856672s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.644819253 +0000 UTC m=+230.315598235" watchObservedRunningTime="2026-03-13 07:42:50.645856672 +0000 UTC m=+230.316635654" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.674969 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.692684 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.192659703 +0000 UTC m=+230.863438685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.700361 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vwq9g" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.711368 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" podStartSLOduration=171.711354516 podStartE2EDuration="2m51.711354516s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.709830441 +0000 UTC m=+230.380609423" watchObservedRunningTime="2026-03-13 07:42:50.711354516 +0000 UTC m=+230.382133498" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.711919 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zhz5h" podStartSLOduration=170.711912612 podStartE2EDuration="2m50.711912612s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.66847021 +0000 UTC m=+230.339249212" watchObservedRunningTime="2026-03-13 07:42:50.711912612 +0000 UTC m=+230.382691594" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.741053 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" podStartSLOduration=170.741012007 podStartE2EDuration="2m50.741012007s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.739630137 +0000 UTC m=+230.410409119" watchObservedRunningTime="2026-03-13 07:42:50.741012007 +0000 UTC m=+230.411790989" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.776448 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" podStartSLOduration=171.776426297 podStartE2EDuration="2m51.776426297s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.771594136 +0000 UTC m=+230.442373138" watchObservedRunningTime="2026-03-13 07:42:50.776426297 +0000 UTC m=+230.447205289" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.779746 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.781600 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.281559685 +0000 UTC m=+230.952338687 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.884086 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.884446 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.384433944 +0000 UTC m=+231.055212926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.943013 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-zk5kl" podStartSLOduration=7.942991826 podStartE2EDuration="7.942991826s" podCreationTimestamp="2026-03-13 07:42:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.814107191 +0000 UTC m=+230.484886183" watchObservedRunningTime="2026-03-13 07:42:50.942991826 +0000 UTC m=+230.613770808" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.946503 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" podStartSLOduration=170.946467716 podStartE2EDuration="2m50.946467716s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:50.944573091 +0000 UTC m=+230.615352073" watchObservedRunningTime="2026-03-13 07:42:50.946467716 +0000 UTC m=+230.617246698" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.956284 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.986208 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.986398 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.486371676 +0000 UTC m=+231.157150658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:50 crc kubenswrapper[4876]: I0313 07:42:50.986647 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:50 crc kubenswrapper[4876]: E0313 07:42:50.986972 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.486958162 +0000 UTC m=+231.157737134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.088346 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.088779 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.58876362 +0000 UTC m=+231.259542602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.189390 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.189752 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.689730823 +0000 UTC m=+231.360509815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.289313 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:51 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:51 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:51 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.289365 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.291094 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.291466 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.791453568 +0000 UTC m=+231.462232540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.347261 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dfkkx" podStartSLOduration=9.347230679 podStartE2EDuration="9.347230679s" podCreationTimestamp="2026-03-13 07:42:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:51.330686209 +0000 UTC m=+231.001465191" watchObservedRunningTime="2026-03-13 07:42:51.347230679 +0000 UTC m=+231.018009661" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.393393 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.394508 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.894493362 +0000 UTC m=+231.565272344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.410548 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47094: no serving certificate available for the kubelet" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.499507 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.499687 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:51.999660897 +0000 UTC m=+231.670439879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.499793 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.500154 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.000140232 +0000 UTC m=+231.670919214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.557022 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47106: no serving certificate available for the kubelet" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.601608 4876 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dnkf8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.601681 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" podUID="f938364d-5a03-47e8-b863-edf8d217770a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.603749 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.604010 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.103996029 +0000 UTC m=+231.774775011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.608340 4876 patch_prober.go:28] interesting pod/console-operator-58897d9998-5rhkc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.608381 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" podUID="945666f1-2703-4fa7-9063-fd49b7189341" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.608940 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" event={"ID":"0445a35a-4d10-444b-99ba-b0ced7f82f09","Type":"ContainerStarted","Data":"26f5cdf044d33b2130f0a782209b4a521de65f923012abcff266610b56c6799a"} Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.618226 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.651040 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-lw2ss" podStartSLOduration=171.650992713 podStartE2EDuration="2m51.650992713s" podCreationTimestamp="2026-03-13 07:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:51.64467655 +0000 UTC m=+231.315455532" watchObservedRunningTime="2026-03-13 07:42:51.650992713 +0000 UTC m=+231.321771695" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.705059 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.707514 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47108: no serving certificate available for the kubelet" Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.760038 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.260019081 +0000 UTC m=+231.930798063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.830369 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.830942 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.330918611 +0000 UTC m=+232.001697613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.831491 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.832051 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.332040704 +0000 UTC m=+232.002819686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.932828 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:51 crc kubenswrapper[4876]: E0313 07:42:51.933193 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.433178862 +0000 UTC m=+232.103957844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:51 crc kubenswrapper[4876]: I0313 07:42:51.972694 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47118: no serving certificate available for the kubelet" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.033961 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.034375 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.534359431 +0000 UTC m=+232.205138413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.093334 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47134: no serving certificate available for the kubelet" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.103718 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dnkf8" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.142990 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.143477 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.64345732 +0000 UTC m=+232.314236302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.234905 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47142: no serving certificate available for the kubelet" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.244515 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.244944 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.744929908 +0000 UTC m=+232.415708900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.300511 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:52 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:52 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:52 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.300588 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.341884 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47148: no serving certificate available for the kubelet" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.347420 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.347760 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.847717655 +0000 UTC m=+232.518496647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.347863 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.348436 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.848424485 +0000 UTC m=+232.519203467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.449104 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.449378 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.949329156 +0000 UTC m=+232.620108138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.449604 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.449876 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:52.949863632 +0000 UTC m=+232.620642614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.456638 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47158: no serving certificate available for the kubelet" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.550556 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.550928 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.050913168 +0000 UTC m=+232.721692150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.642021 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" event={"ID":"aa7285c8-4679-495d-afca-6d15a19cfa4c","Type":"ContainerStarted","Data":"61f13c61246b04e113c6aa27e5acce26063cc18455d2df3a9c0a231a7c7cc6a0"} Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.652702 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.653221 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.153203939 +0000 UTC m=+232.823982921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.753852 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.755175 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.255160012 +0000 UTC m=+232.925938994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.773623 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kgzdt"] Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.774765 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: W0313 07:42:52.778308 4876 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.778362 4876 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.855345 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl8wt\" (UniqueName: \"kubernetes.io/projected/af6d5722-1354-4021-9ba4-6ad9a1286c60-kube-api-access-zl8wt\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.855400 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.855516 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-utilities\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.855589 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-catalog-content\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.855712 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.355697163 +0000 UTC m=+233.026476145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.956488 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.956762 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-utilities\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.956890 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-catalog-content\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.956936 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl8wt\" (UniqueName: \"kubernetes.io/projected/af6d5722-1354-4021-9ba4-6ad9a1286c60-kube-api-access-zl8wt\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: E0313 07:42:52.957335 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.457318915 +0000 UTC m=+233.128097897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.957784 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-utilities\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.958285 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-catalog-content\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.965722 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fmcl5"] Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.967102 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:52 crc kubenswrapper[4876]: I0313 07:42:52.973860 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.062404 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ld7c\" (UniqueName: \"kubernetes.io/projected/8c486d8f-87cf-42c0-9b8c-8f715755edc8-kube-api-access-9ld7c\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.062458 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-utilities\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.062562 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.062630 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-catalog-content\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.066702 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.566684432 +0000 UTC m=+233.237463414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.085304 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fmcl5"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.086599 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kgzdt"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.106461 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl8wt\" (UniqueName: \"kubernetes.io/projected/af6d5722-1354-4021-9ba4-6ad9a1286c60-kube-api-access-zl8wt\") pod \"certified-operators-kgzdt\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.170834 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cf5vz"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.171934 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.177401 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.177693 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-catalog-content\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.177720 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ld7c\" (UniqueName: \"kubernetes.io/projected/8c486d8f-87cf-42c0-9b8c-8f715755edc8-kube-api-access-9ld7c\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.177742 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-utilities\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.178147 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-utilities\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.178290 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.678223052 +0000 UTC m=+233.349002034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.178503 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-catalog-content\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.208829 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cf5vz"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.261717 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ld7c\" (UniqueName: \"kubernetes.io/projected/8c486d8f-87cf-42c0-9b8c-8f715755edc8-kube-api-access-9ld7c\") pod \"community-operators-fmcl5\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.282216 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-catalog-content\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.282402 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-utilities\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.282458 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr79z\" (UniqueName: \"kubernetes.io/projected/ddbd46d5-05cd-4d71-b085-142b13687f63-kube-api-access-rr79z\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.282494 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.282835 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.782818281 +0000 UTC m=+233.453597273 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.287207 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47160: no serving certificate available for the kubelet" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.290125 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.290576 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.290762 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:53 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:53 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:53 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.290798 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.291230 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.299359 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7f5nt" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.303902 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.304376 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.315338 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.346287 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mtcn4"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.347593 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.385500 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtcn4"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.389830 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.390056 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-utilities\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.390080 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.890053486 +0000 UTC m=+233.560832468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.390120 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr79z\" (UniqueName: \"kubernetes.io/projected/ddbd46d5-05cd-4d71-b085-142b13687f63-kube-api-access-rr79z\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.390178 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.390225 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-catalog-content\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.390275 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d3fa307-6dbc-4387-ab37-903b95f39b93-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1d3fa307-6dbc-4387-ab37-903b95f39b93\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.390290 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d3fa307-6dbc-4387-ab37-903b95f39b93-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1d3fa307-6dbc-4387-ab37-903b95f39b93\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.390538 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-utilities\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.390952 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.890939482 +0000 UTC m=+233.561718464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.397391 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-catalog-content\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.447983 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr79z\" (UniqueName: \"kubernetes.io/projected/ddbd46d5-05cd-4d71-b085-142b13687f63-kube-api-access-rr79z\") pod \"certified-operators-cf5vz\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.490964 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.491291 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-catalog-content\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.491337 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrvnp\" (UniqueName: \"kubernetes.io/projected/18c1ffa5-3580-48f1-8df2-5f442d47062d-kube-api-access-zrvnp\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.491431 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-utilities\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.491451 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d3fa307-6dbc-4387-ab37-903b95f39b93-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1d3fa307-6dbc-4387-ab37-903b95f39b93\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.491466 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d3fa307-6dbc-4387-ab37-903b95f39b93-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1d3fa307-6dbc-4387-ab37-903b95f39b93\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.491824 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:53.991811022 +0000 UTC m=+233.662590004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.491869 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d3fa307-6dbc-4387-ab37-903b95f39b93-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1d3fa307-6dbc-4387-ab37-903b95f39b93\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.574614 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d3fa307-6dbc-4387-ab37-903b95f39b93-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1d3fa307-6dbc-4387-ab37-903b95f39b93\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.595559 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-catalog-content\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.595604 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrvnp\" (UniqueName: \"kubernetes.io/projected/18c1ffa5-3580-48f1-8df2-5f442d47062d-kube-api-access-zrvnp\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.595676 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.595703 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-utilities\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.596097 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-utilities\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.596344 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-catalog-content\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.596699 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.096689019 +0000 UTC m=+233.767468001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.627430 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.665118 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrvnp\" (UniqueName: \"kubernetes.io/projected/18c1ffa5-3580-48f1-8df2-5f442d47062d-kube-api-access-zrvnp\") pod \"community-operators-mtcn4\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.696257 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.696361 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.196337364 +0000 UTC m=+233.867116356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.696474 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.696799 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.196787667 +0000 UTC m=+233.867566649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.696956 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" event={"ID":"aa7285c8-4679-495d-afca-6d15a19cfa4c","Type":"ContainerStarted","Data":"ea769447898db482b7f3e9ad751a6d39b1d40e35c94e971efad4c234a2b1ea0d"} Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.723659 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.797751 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.797937 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.297915265 +0000 UTC m=+233.968694247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.798282 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.799139 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.29912847 +0000 UTC m=+233.969907442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.871934 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.872585 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.876692 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.876872 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.899372 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:53 crc kubenswrapper[4876]: E0313 07:42:53.899700 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.399686311 +0000 UTC m=+234.070465293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:53 crc kubenswrapper[4876]: I0313 07:42:53.905828 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.002583 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1578f9c7-03ac-45ac-92d6-cc655178e60f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1578f9c7-03ac-45ac-92d6-cc655178e60f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.002639 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.002670 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1578f9c7-03ac-45ac-92d6-cc655178e60f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1578f9c7-03ac-45ac-92d6-cc655178e60f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:54 crc kubenswrapper[4876]: E0313 07:42:54.003051 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.503037884 +0000 UTC m=+234.173816866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.025374 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nkhkk"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.025582 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" podUID="ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" containerName="controller-manager" containerID="cri-o://cc6ed110258c6762ee845c2f13301516f9c7b773ae67bc9da6bd7e59046c69ce" gracePeriod=30 Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.105607 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.105764 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1578f9c7-03ac-45ac-92d6-cc655178e60f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1578f9c7-03ac-45ac-92d6-cc655178e60f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.105873 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1578f9c7-03ac-45ac-92d6-cc655178e60f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1578f9c7-03ac-45ac-92d6-cc655178e60f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.105945 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1578f9c7-03ac-45ac-92d6-cc655178e60f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1578f9c7-03ac-45ac-92d6-cc655178e60f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:54 crc kubenswrapper[4876]: E0313 07:42:54.106020 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.605997546 +0000 UTC m=+234.276776528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.136820 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1578f9c7-03ac-45ac-92d6-cc655178e60f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1578f9c7-03ac-45ac-92d6-cc655178e60f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.137021 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.137269 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" podUID="f8b52e5e-95cb-4d81-b3c2-d014955b31fd" containerName="route-controller-manager" containerID="cri-o://c0dd9b9df06a80699c8834997b70e44c27af2e486bb85af9a12062f486f3abb7" gracePeriod=30 Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.207302 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:54 crc kubenswrapper[4876]: E0313 07:42:54.207698 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 07:42:54.707684149 +0000 UTC m=+234.378463131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c5pvl" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.211958 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.248738 4876 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.281774 4876 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-13T07:42:54.248776864Z","Handler":null,"Name":""} Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.286746 4876 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.286786 4876 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.287313 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:54 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:54 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:54 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.288068 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.308410 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.311063 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.321135 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.321247 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.389090 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.409460 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.432362 4876 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.432414 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.477443 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.494004 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c5pvl\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.608215 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.608535 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.643556 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.645208 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47164: no serving certificate available for the kubelet" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.648394 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtcn4"] Mar 13 07:42:54 crc kubenswrapper[4876]: W0313 07:42:54.683701 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18c1ffa5_3580_48f1_8df2_5f442d47062d.slice/crio-fcd55685e2d927ed8f2ddbb66a591fe88507a2d538d334fa47c422fcf5092a2f WatchSource:0}: Error finding container fcd55685e2d927ed8f2ddbb66a591fe88507a2d538d334fa47c422fcf5092a2f: Status 404 returned error can't find the container with id fcd55685e2d927ed8f2ddbb66a591fe88507a2d538d334fa47c422fcf5092a2f Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.714219 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.732653 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fmcl5"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.744573 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1d3fa307-6dbc-4387-ab37-903b95f39b93","Type":"ContainerStarted","Data":"1e1eda3da4bdae9333b3de9f10f3a96f7d378dcef2751fc8a377783e5ec61f0e"} Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.746030 4876 generic.go:334] "Generic (PLEG): container finished" podID="ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" containerID="cc6ed110258c6762ee845c2f13301516f9c7b773ae67bc9da6bd7e59046c69ce" exitCode=0 Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.746081 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" event={"ID":"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2","Type":"ContainerDied","Data":"cc6ed110258c6762ee845c2f13301516f9c7b773ae67bc9da6bd7e59046c69ce"} Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.752285 4876 generic.go:334] "Generic (PLEG): container finished" podID="f8b52e5e-95cb-4d81-b3c2-d014955b31fd" containerID="c0dd9b9df06a80699c8834997b70e44c27af2e486bb85af9a12062f486f3abb7" exitCode=0 Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.752371 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" event={"ID":"f8b52e5e-95cb-4d81-b3c2-d014955b31fd","Type":"ContainerDied","Data":"c0dd9b9df06a80699c8834997b70e44c27af2e486bb85af9a12062f486f3abb7"} Mar 13 07:42:54 crc kubenswrapper[4876]: W0313 07:42:54.767770 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c486d8f_87cf_42c0_9b8c_8f715755edc8.slice/crio-6bfe74f805be6ad373fc748949952661b08729715e854d215c05420ca1ee568f WatchSource:0}: Error finding container 6bfe74f805be6ad373fc748949952661b08729715e854d215c05420ca1ee568f: Status 404 returned error can't find the container with id 6bfe74f805be6ad373fc748949952661b08729715e854d215c05420ca1ee568f Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.775068 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" event={"ID":"aa7285c8-4679-495d-afca-6d15a19cfa4c","Type":"ContainerStarted","Data":"2affb79be5409d787e8dca92ad8d0c50a6ec314789352b71e04914035b2654b8"} Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.791327 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtcn4" event={"ID":"18c1ffa5-3580-48f1-8df2-5f442d47062d","Type":"ContainerStarted","Data":"fcd55685e2d927ed8f2ddbb66a591fe88507a2d538d334fa47c422fcf5092a2f"} Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.904424 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-m2gxf" podStartSLOduration=12.90440553 podStartE2EDuration="12.90440553s" podCreationTimestamp="2026-03-13 07:42:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:54.805431375 +0000 UTC m=+234.476210347" watchObservedRunningTime="2026-03-13 07:42:54.90440553 +0000 UTC m=+234.575184512" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.907284 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2q97b"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.910517 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.912330 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.917140 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q97b"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.930919 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kgzdt"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.960900 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cf5vz"] Mar 13 07:42:54 crc kubenswrapper[4876]: I0313 07:42:54.988869 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.018717 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.021563 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-utilities\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.021590 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-catalog-content\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.021785 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws8qb\" (UniqueName: \"kubernetes.io/projected/63f3a5db-0db1-4486-a982-39c60bda9dc4-kube-api-access-ws8qb\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.032793 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.092335 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.116162 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5pvl"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123538 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-config\") pod \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123582 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-config\") pod \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123614 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-serving-cert\") pod \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123646 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-serving-cert\") pod \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123664 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-client-ca\") pod \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123682 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-proxy-ca-bundles\") pod \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123716 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cttzt\" (UniqueName: \"kubernetes.io/projected/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-kube-api-access-cttzt\") pod \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\" (UID: \"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123749 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-client-ca\") pod \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123796 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n6t6\" (UniqueName: \"kubernetes.io/projected/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-kube-api-access-4n6t6\") pod \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\" (UID: \"f8b52e5e-95cb-4d81-b3c2-d014955b31fd\") " Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.123981 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws8qb\" (UniqueName: \"kubernetes.io/projected/63f3a5db-0db1-4486-a982-39c60bda9dc4-kube-api-access-ws8qb\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.124041 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-utilities\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.124058 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-catalog-content\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.124886 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-client-ca" (OuterVolumeSpecName: "client-ca") pod "f8b52e5e-95cb-4d81-b3c2-d014955b31fd" (UID: "f8b52e5e-95cb-4d81-b3c2-d014955b31fd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.125162 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-config" (OuterVolumeSpecName: "config") pod "f8b52e5e-95cb-4d81-b3c2-d014955b31fd" (UID: "f8b52e5e-95cb-4d81-b3c2-d014955b31fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.125511 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-client-ca" (OuterVolumeSpecName: "client-ca") pod "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" (UID: "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.125604 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" (UID: "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.125629 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-utilities\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.125773 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-catalog-content\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.126296 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-config" (OuterVolumeSpecName: "config") pod "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" (UID: "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.130621 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.131156 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" (UID: "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.131406 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.131545 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f8b52e5e-95cb-4d81-b3c2-d014955b31fd" (UID: "f8b52e5e-95cb-4d81-b3c2-d014955b31fd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.143868 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.145227 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.145478 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-kube-api-access-4n6t6" (OuterVolumeSpecName: "kube-api-access-4n6t6") pod "f8b52e5e-95cb-4d81-b3c2-d014955b31fd" (UID: "f8b52e5e-95cb-4d81-b3c2-d014955b31fd"). InnerVolumeSpecName "kube-api-access-4n6t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.147074 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws8qb\" (UniqueName: \"kubernetes.io/projected/63f3a5db-0db1-4486-a982-39c60bda9dc4-kube-api-access-ws8qb\") pod \"redhat-marketplace-2q97b\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.150312 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-kube-api-access-cttzt" (OuterVolumeSpecName: "kube-api-access-cttzt") pod "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" (UID: "ad2feea5-fef1-4fcf-9cb3-7d281559d1b2"). InnerVolumeSpecName "kube-api-access-cttzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.179416 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.180544 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229365 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229391 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229402 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229414 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cttzt\" (UniqueName: \"kubernetes.io/projected/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-kube-api-access-cttzt\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229426 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229436 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n6t6\" (UniqueName: \"kubernetes.io/projected/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-kube-api-access-4n6t6\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229447 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229457 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8b52e5e-95cb-4d81-b3c2-d014955b31fd-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.229466 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.235368 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.235764 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.237964 4876 patch_prober.go:28] interesting pod/console-f9d7485db-g65sf container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.238010 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-g65sf" podUID="62fdbab2-2be8-42cc-a936-a87ffd085a0e" containerName="console" probeResult="failure" output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.282199 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.286469 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:55 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:55 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:55 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.286580 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.295224 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bg8rq"] Mar 13 07:42:55 crc kubenswrapper[4876]: E0313 07:42:55.295497 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" containerName="controller-manager" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.295516 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" containerName="controller-manager" Mar 13 07:42:55 crc kubenswrapper[4876]: E0313 07:42:55.295533 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b52e5e-95cb-4d81-b3c2-d014955b31fd" containerName="route-controller-manager" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.295541 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b52e5e-95cb-4d81-b3c2-d014955b31fd" containerName="route-controller-manager" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.295654 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8b52e5e-95cb-4d81-b3c2-d014955b31fd" containerName="route-controller-manager" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.295676 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" containerName="controller-manager" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.296696 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.302187 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.302273 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.302868 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.302893 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.310594 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bg8rq"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.372791 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.438895 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-utilities\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.438940 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5lkh\" (UniqueName: \"kubernetes.io/projected/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-kube-api-access-r5lkh\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.439043 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-catalog-content\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.540729 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-utilities\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.541108 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5lkh\" (UniqueName: \"kubernetes.io/projected/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-kube-api-access-r5lkh\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.541179 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-catalog-content\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.541884 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-catalog-content\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.542149 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-utilities\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.562704 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5lkh\" (UniqueName: \"kubernetes.io/projected/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-kube-api-access-r5lkh\") pod \"redhat-marketplace-bg8rq\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.581056 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-5rhkc" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.805448 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.815897 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.816455 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nkhkk" event={"ID":"ad2feea5-fef1-4fcf-9cb3-7d281559d1b2","Type":"ContainerDied","Data":"220199b0d93904fe88f39ad029661bfb85b3c17beaf3cafd29e88a02638a2d9d"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.816498 4876 scope.go:117] "RemoveContainer" containerID="cc6ed110258c6762ee845c2f13301516f9c7b773ae67bc9da6bd7e59046c69ce" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.829606 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q97b"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.829655 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" event={"ID":"f8b52e5e-95cb-4d81-b3c2-d014955b31fd","Type":"ContainerDied","Data":"fd4d3b5581750eb5c9411366d6355ea96e64f230ff0f8c6da5ed255d457b8f00"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.829722 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.842374 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1578f9c7-03ac-45ac-92d6-cc655178e60f","Type":"ContainerStarted","Data":"a0f1fc1a2406478d9214a24301ff7b7662c20e62e42e40e8d30fd89943f8e60e"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.842617 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1578f9c7-03ac-45ac-92d6-cc655178e60f","Type":"ContainerStarted","Data":"44ade10794c51587f71379e297e90815e10f562d7ccd77ba6781d40c21c946fe"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.846404 4876 generic.go:334] "Generic (PLEG): container finished" podID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerID="3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33" exitCode=0 Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.846577 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cf5vz" event={"ID":"ddbd46d5-05cd-4d71-b085-142b13687f63","Type":"ContainerDied","Data":"3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.846695 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cf5vz" event={"ID":"ddbd46d5-05cd-4d71-b085-142b13687f63","Type":"ContainerStarted","Data":"59d822d7a5fdea9cd1e2e79fa2183c85fea8e238f88747f59d4dc07b60226db5"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.849016 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" event={"ID":"0991deda-4e1a-4fe7-92e9-01bbaf54efb0","Type":"ContainerStarted","Data":"2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.849068 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" event={"ID":"0991deda-4e1a-4fe7-92e9-01bbaf54efb0","Type":"ContainerStarted","Data":"25060b51be92f6e5fb37c755cd1c6aebc9008c6acca9803c25642996b49d3e5a"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.849903 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.850928 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1d3fa307-6dbc-4387-ab37-903b95f39b93","Type":"ContainerStarted","Data":"3cd54fc34a9c9ff04c1f2efb603253bed7e49b2d7f265304701f5dd1e1437d09"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.853297 4876 generic.go:334] "Generic (PLEG): container finished" podID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerID="ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc" exitCode=0 Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.853387 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgzdt" event={"ID":"af6d5722-1354-4021-9ba4-6ad9a1286c60","Type":"ContainerDied","Data":"ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.853411 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgzdt" event={"ID":"af6d5722-1354-4021-9ba4-6ad9a1286c60","Type":"ContainerStarted","Data":"486d7fdf7bb9b66bb0104a13afa68c6e8110ff958605093e669b58880a6d6c00"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.854559 4876 generic.go:334] "Generic (PLEG): container finished" podID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerID="44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24" exitCode=0 Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.854604 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcl5" event={"ID":"8c486d8f-87cf-42c0-9b8c-8f715755edc8","Type":"ContainerDied","Data":"44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.854622 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcl5" event={"ID":"8c486d8f-87cf-42c0-9b8c-8f715755edc8","Type":"ContainerStarted","Data":"6bfe74f805be6ad373fc748949952661b08729715e854d215c05420ca1ee568f"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.855650 4876 scope.go:117] "RemoveContainer" containerID="c0dd9b9df06a80699c8834997b70e44c27af2e486bb85af9a12062f486f3abb7" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.856756 4876 generic.go:334] "Generic (PLEG): container finished" podID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerID="4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94" exitCode=0 Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.858417 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtcn4" event={"ID":"18c1ffa5-3580-48f1-8df2-5f442d47062d","Type":"ContainerDied","Data":"4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94"} Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.872618 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.872594698 podStartE2EDuration="2.872594698s" podCreationTimestamp="2026-03-13 07:42:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:55.869088386 +0000 UTC m=+235.539867368" watchObservedRunningTime="2026-03-13 07:42:55.872594698 +0000 UTC m=+235.543373680" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.874781 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6qqdn" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.884844 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xttrq" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.891424 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.897168 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jdqxn"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.916578 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kdxtz"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.917829 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.920994 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" podStartSLOduration=176.920980514 podStartE2EDuration="2m56.920980514s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:55.905886335 +0000 UTC m=+235.576665327" watchObservedRunningTime="2026-03-13 07:42:55.920980514 +0000 UTC m=+235.591759496" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.921450 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.925785 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kdxtz"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.933853 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-64cd5df9d4-kt45r"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.950795 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-679d787886-82v44"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.951274 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64cd5df9d4-kt45r"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.951356 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.951814 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.952918 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-679d787886-82v44"] Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.954955 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.957719 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.958136 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.958320 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.958402 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.958599 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.958704 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.958747 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.958898 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.959292 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.959436 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.960207 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 07:42:55 crc kubenswrapper[4876]: I0313 07:42:55.968444 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.015741 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nkhkk"] Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.018707 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nkhkk"] Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.024102 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.02408758 podStartE2EDuration="3.02408758s" podCreationTimestamp="2026-03-13 07:42:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:56.023105801 +0000 UTC m=+235.693884783" watchObservedRunningTime="2026-03-13 07:42:56.02408758 +0000 UTC m=+235.694866562" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.064905 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-client-ca\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.064943 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-client-ca\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.064965 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-catalog-content\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.064990 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h95dx\" (UniqueName: \"kubernetes.io/projected/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-kube-api-access-h95dx\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.065039 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xxhq\" (UniqueName: \"kubernetes.io/projected/53894f81-37aa-4a7e-837b-0f5e722352af-kube-api-access-4xxhq\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.065077 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53894f81-37aa-4a7e-837b-0f5e722352af-serving-cert\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.065102 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hqlf\" (UniqueName: \"kubernetes.io/projected/b694aa35-ed97-4c15-9161-8deaf40656f6-kube-api-access-5hqlf\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.065123 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-config\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.065141 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-utilities\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.065169 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b694aa35-ed97-4c15-9161-8deaf40656f6-serving-cert\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.065190 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-config\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.065252 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-proxy-ca-bundles\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.080550 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167212 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-proxy-ca-bundles\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167277 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-client-ca\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167304 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-client-ca\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167341 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-catalog-content\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167369 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h95dx\" (UniqueName: \"kubernetes.io/projected/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-kube-api-access-h95dx\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167419 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xxhq\" (UniqueName: \"kubernetes.io/projected/53894f81-37aa-4a7e-837b-0f5e722352af-kube-api-access-4xxhq\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167474 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53894f81-37aa-4a7e-837b-0f5e722352af-serving-cert\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167518 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hqlf\" (UniqueName: \"kubernetes.io/projected/b694aa35-ed97-4c15-9161-8deaf40656f6-kube-api-access-5hqlf\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167559 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-config\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167598 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-utilities\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167621 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b694aa35-ed97-4c15-9161-8deaf40656f6-serving-cert\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.167650 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-config\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.170388 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-client-ca\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.171510 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-proxy-ca-bundles\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.172309 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-config\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.173146 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-utilities\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.173862 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-config\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.174389 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-catalog-content\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.182187 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-client-ca\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.187215 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b694aa35-ed97-4c15-9161-8deaf40656f6-serving-cert\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.187937 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53894f81-37aa-4a7e-837b-0f5e722352af-serving-cert\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.241032 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xxhq\" (UniqueName: \"kubernetes.io/projected/53894f81-37aa-4a7e-837b-0f5e722352af-kube-api-access-4xxhq\") pod \"controller-manager-64cd5df9d4-kt45r\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.247336 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h95dx\" (UniqueName: \"kubernetes.io/projected/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-kube-api-access-h95dx\") pod \"redhat-operators-kdxtz\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.268392 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hqlf\" (UniqueName: \"kubernetes.io/projected/b694aa35-ed97-4c15-9161-8deaf40656f6-kube-api-access-5hqlf\") pod \"route-controller-manager-679d787886-82v44\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.268711 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.288448 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:56 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:56 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:56 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.288498 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.314371 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gkxcc"] Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.316106 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.333938 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gkxcc"] Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.372122 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7bxl\" (UniqueName: \"kubernetes.io/projected/68fd169c-1d47-4f5b-a528-201a1ea159a1-kube-api-access-v7bxl\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.372204 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-catalog-content\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.372327 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-utilities\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.395289 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bg8rq"] Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.474174 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.474674 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-utilities\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.474713 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7bxl\" (UniqueName: \"kubernetes.io/projected/68fd169c-1d47-4f5b-a528-201a1ea159a1-kube-api-access-v7bxl\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.474755 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-catalog-content\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.475465 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-utilities\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.475508 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-catalog-content\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.492462 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.496801 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7bxl\" (UniqueName: \"kubernetes.io/projected/68fd169c-1d47-4f5b-a528-201a1ea159a1-kube-api-access-v7bxl\") pod \"redhat-operators-gkxcc\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.648083 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.832869 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64cd5df9d4-kt45r"] Mar 13 07:42:56 crc kubenswrapper[4876]: W0313 07:42:56.864868 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53894f81_37aa_4a7e_837b_0f5e722352af.slice/crio-12d09b150a43a7ba5e5e13ced555d4506b54d9ea96d3af93d7ee9f5e2ab7129c WatchSource:0}: Error finding container 12d09b150a43a7ba5e5e13ced555d4506b54d9ea96d3af93d7ee9f5e2ab7129c: Status 404 returned error can't find the container with id 12d09b150a43a7ba5e5e13ced555d4506b54d9ea96d3af93d7ee9f5e2ab7129c Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.946059 4876 generic.go:334] "Generic (PLEG): container finished" podID="1d3fa307-6dbc-4387-ab37-903b95f39b93" containerID="3cd54fc34a9c9ff04c1f2efb603253bed7e49b2d7f265304701f5dd1e1437d09" exitCode=0 Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.946168 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1d3fa307-6dbc-4387-ab37-903b95f39b93","Type":"ContainerDied","Data":"3cd54fc34a9c9ff04c1f2efb603253bed7e49b2d7f265304701f5dd1e1437d09"} Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.952666 4876 generic.go:334] "Generic (PLEG): container finished" podID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerID="c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2" exitCode=0 Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.952735 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q97b" event={"ID":"63f3a5db-0db1-4486-a982-39c60bda9dc4","Type":"ContainerDied","Data":"c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2"} Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.952762 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q97b" event={"ID":"63f3a5db-0db1-4486-a982-39c60bda9dc4","Type":"ContainerStarted","Data":"521b126297d22b9a1389665d013f1e667e8f021cdccbbbc29b37c570afdab435"} Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.955930 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" event={"ID":"53894f81-37aa-4a7e-837b-0f5e722352af","Type":"ContainerStarted","Data":"12d09b150a43a7ba5e5e13ced555d4506b54d9ea96d3af93d7ee9f5e2ab7129c"} Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.973153 4876 generic.go:334] "Generic (PLEG): container finished" podID="1578f9c7-03ac-45ac-92d6-cc655178e60f" containerID="a0f1fc1a2406478d9214a24301ff7b7662c20e62e42e40e8d30fd89943f8e60e" exitCode=0 Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.973224 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1578f9c7-03ac-45ac-92d6-cc655178e60f","Type":"ContainerDied","Data":"a0f1fc1a2406478d9214a24301ff7b7662c20e62e42e40e8d30fd89943f8e60e"} Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.983605 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kdxtz"] Mar 13 07:42:56 crc kubenswrapper[4876]: I0313 07:42:56.993042 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bg8rq" event={"ID":"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1","Type":"ContainerStarted","Data":"faaa89fa9f550f216225d2c0149e2ff2ec70f09230a7f549797faddfb647c1b4"} Mar 13 07:42:57 crc kubenswrapper[4876]: W0313 07:42:56.997323 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7fbdc82_2033_4493_8f3a_ee17b88eb94e.slice/crio-e272a0aa889f2f6688e65c378f1e24df44a9de205b08e05e5972dd74aac96bbc WatchSource:0}: Error finding container e272a0aa889f2f6688e65c378f1e24df44a9de205b08e05e5972dd74aac96bbc: Status 404 returned error can't find the container with id e272a0aa889f2f6688e65c378f1e24df44a9de205b08e05e5972dd74aac96bbc Mar 13 07:42:57 crc kubenswrapper[4876]: I0313 07:42:57.046262 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad2feea5-fef1-4fcf-9cb3-7d281559d1b2" path="/var/lib/kubelet/pods/ad2feea5-fef1-4fcf-9cb3-7d281559d1b2/volumes" Mar 13 07:42:57 crc kubenswrapper[4876]: I0313 07:42:57.047059 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8b52e5e-95cb-4d81-b3c2-d014955b31fd" path="/var/lib/kubelet/pods/f8b52e5e-95cb-4d81-b3c2-d014955b31fd/volumes" Mar 13 07:42:57 crc kubenswrapper[4876]: I0313 07:42:57.128635 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-679d787886-82v44"] Mar 13 07:42:57 crc kubenswrapper[4876]: W0313 07:42:57.230293 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb694aa35_ed97_4c15_9161_8deaf40656f6.slice/crio-4d446e976750c4d5bd1558b023f74ee2b6b61ab2f9cbe6c312341b1606b5bcc4 WatchSource:0}: Error finding container 4d446e976750c4d5bd1558b023f74ee2b6b61ab2f9cbe6c312341b1606b5bcc4: Status 404 returned error can't find the container with id 4d446e976750c4d5bd1558b023f74ee2b6b61ab2f9cbe6c312341b1606b5bcc4 Mar 13 07:42:57 crc kubenswrapper[4876]: I0313 07:42:57.256638 4876 ???:1] "http: TLS handshake error from 192.168.126.11:47166: no serving certificate available for the kubelet" Mar 13 07:42:57 crc kubenswrapper[4876]: I0313 07:42:57.290229 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:57 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:57 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:57 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:57 crc kubenswrapper[4876]: I0313 07:42:57.290359 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:57 crc kubenswrapper[4876]: I0313 07:42:57.373394 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gkxcc"] Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.049985 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" event={"ID":"53894f81-37aa-4a7e-837b-0f5e722352af","Type":"ContainerStarted","Data":"417b3d65fa933dbef0d5c6eb42b93643e7d0d0e51d81c6197403aa26835e47e3"} Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.050887 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.066557 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.068922 4876 generic.go:334] "Generic (PLEG): container finished" podID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerID="bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137" exitCode=0 Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.069018 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gkxcc" event={"ID":"68fd169c-1d47-4f5b-a528-201a1ea159a1","Type":"ContainerDied","Data":"bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137"} Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.069073 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gkxcc" event={"ID":"68fd169c-1d47-4f5b-a528-201a1ea159a1","Type":"ContainerStarted","Data":"a8a2f6aadef356e01fcb36bd4d8bd6294768da7386fbf3a366f125291e44d7f0"} Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.070814 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" podStartSLOduration=4.0707991 podStartE2EDuration="4.0707991s" podCreationTimestamp="2026-03-13 07:42:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:58.070352637 +0000 UTC m=+237.741131629" watchObservedRunningTime="2026-03-13 07:42:58.0707991 +0000 UTC m=+237.741578082" Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.095715 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" event={"ID":"b694aa35-ed97-4c15-9161-8deaf40656f6","Type":"ContainerStarted","Data":"9fa46fab2dc7890152379da63bb9d40e3ed459ce381562a9736403541ffb4d9b"} Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.095799 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" event={"ID":"b694aa35-ed97-4c15-9161-8deaf40656f6","Type":"ContainerStarted","Data":"4d446e976750c4d5bd1558b023f74ee2b6b61ab2f9cbe6c312341b1606b5bcc4"} Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.096679 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.118391 4876 generic.go:334] "Generic (PLEG): container finished" podID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerID="72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44" exitCode=0 Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.119204 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bg8rq" event={"ID":"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1","Type":"ContainerDied","Data":"72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44"} Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.131959 4876 generic.go:334] "Generic (PLEG): container finished" podID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerID="aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62" exitCode=0 Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.133118 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdxtz" event={"ID":"f7fbdc82-2033-4493-8f3a-ee17b88eb94e","Type":"ContainerDied","Data":"aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62"} Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.133141 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdxtz" event={"ID":"f7fbdc82-2033-4493-8f3a-ee17b88eb94e","Type":"ContainerStarted","Data":"e272a0aa889f2f6688e65c378f1e24df44a9de205b08e05e5972dd74aac96bbc"} Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.272786 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" podStartSLOduration=4.272767998 podStartE2EDuration="4.272767998s" podCreationTimestamp="2026-03-13 07:42:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:42:58.266097444 +0000 UTC m=+237.936876426" watchObservedRunningTime="2026-03-13 07:42:58.272767998 +0000 UTC m=+237.943546980" Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.301845 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:58 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:58 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:58 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.301918 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.304106 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.463733 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dfkkx" Mar 13 07:42:58 crc kubenswrapper[4876]: I0313 07:42:58.969696 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.032373 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.073655 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d3fa307-6dbc-4387-ab37-903b95f39b93-kubelet-dir\") pod \"1d3fa307-6dbc-4387-ab37-903b95f39b93\" (UID: \"1d3fa307-6dbc-4387-ab37-903b95f39b93\") " Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.073706 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1578f9c7-03ac-45ac-92d6-cc655178e60f-kubelet-dir\") pod \"1578f9c7-03ac-45ac-92d6-cc655178e60f\" (UID: \"1578f9c7-03ac-45ac-92d6-cc655178e60f\") " Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.073866 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d3fa307-6dbc-4387-ab37-903b95f39b93-kube-api-access\") pod \"1d3fa307-6dbc-4387-ab37-903b95f39b93\" (UID: \"1d3fa307-6dbc-4387-ab37-903b95f39b93\") " Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.073902 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1578f9c7-03ac-45ac-92d6-cc655178e60f-kube-api-access\") pod \"1578f9c7-03ac-45ac-92d6-cc655178e60f\" (UID: \"1578f9c7-03ac-45ac-92d6-cc655178e60f\") " Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.076198 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d3fa307-6dbc-4387-ab37-903b95f39b93-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1d3fa307-6dbc-4387-ab37-903b95f39b93" (UID: "1d3fa307-6dbc-4387-ab37-903b95f39b93"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.076429 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1578f9c7-03ac-45ac-92d6-cc655178e60f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1578f9c7-03ac-45ac-92d6-cc655178e60f" (UID: "1578f9c7-03ac-45ac-92d6-cc655178e60f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.087397 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1578f9c7-03ac-45ac-92d6-cc655178e60f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1578f9c7-03ac-45ac-92d6-cc655178e60f" (UID: "1578f9c7-03ac-45ac-92d6-cc655178e60f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.115754 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d3fa307-6dbc-4387-ab37-903b95f39b93-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1d3fa307-6dbc-4387-ab37-903b95f39b93" (UID: "1d3fa307-6dbc-4387-ab37-903b95f39b93"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.175540 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d3fa307-6dbc-4387-ab37-903b95f39b93-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.175575 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1578f9c7-03ac-45ac-92d6-cc655178e60f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.175586 4876 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d3fa307-6dbc-4387-ab37-903b95f39b93-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.175594 4876 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1578f9c7-03ac-45ac-92d6-cc655178e60f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.251684 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1578f9c7-03ac-45ac-92d6-cc655178e60f","Type":"ContainerDied","Data":"44ade10794c51587f71379e297e90815e10f562d7ccd77ba6781d40c21c946fe"} Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.251733 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44ade10794c51587f71379e297e90815e10f562d7ccd77ba6781d40c21c946fe" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.251815 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.286457 4876 generic.go:334] "Generic (PLEG): container finished" podID="efc8b898-1330-4381-b780-ee41a9514f05" containerID="181d010f2fcc26b78278f10c1875f24e976a91b573cefd134991bf5558ea2faa" exitCode=0 Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.286556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" event={"ID":"efc8b898-1330-4381-b780-ee41a9514f05","Type":"ContainerDied","Data":"181d010f2fcc26b78278f10c1875f24e976a91b573cefd134991bf5558ea2faa"} Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.298765 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1d3fa307-6dbc-4387-ab37-903b95f39b93","Type":"ContainerDied","Data":"1e1eda3da4bdae9333b3de9f10f3a96f7d378dcef2751fc8a377783e5ec61f0e"} Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.298844 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e1eda3da4bdae9333b3de9f10f3a96f7d378dcef2751fc8a377783e5ec61f0e" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.299421 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.327796 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:42:59 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:42:59 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:42:59 crc kubenswrapper[4876]: healthz check failed Mar 13 07:42:59 crc kubenswrapper[4876]: I0313 07:42:59.327840 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.285806 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:43:00 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:43:00 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:43:00 crc kubenswrapper[4876]: healthz check failed Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.287564 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.871198 4876 ???:1] "http: TLS handshake error from 192.168.126.11:54690: no serving certificate available for the kubelet" Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.884180 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.931027 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnljr\" (UniqueName: \"kubernetes.io/projected/efc8b898-1330-4381-b780-ee41a9514f05-kube-api-access-bnljr\") pod \"efc8b898-1330-4381-b780-ee41a9514f05\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.931262 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc8b898-1330-4381-b780-ee41a9514f05-config-volume\") pod \"efc8b898-1330-4381-b780-ee41a9514f05\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.931396 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efc8b898-1330-4381-b780-ee41a9514f05-secret-volume\") pod \"efc8b898-1330-4381-b780-ee41a9514f05\" (UID: \"efc8b898-1330-4381-b780-ee41a9514f05\") " Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.932759 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efc8b898-1330-4381-b780-ee41a9514f05-config-volume" (OuterVolumeSpecName: "config-volume") pod "efc8b898-1330-4381-b780-ee41a9514f05" (UID: "efc8b898-1330-4381-b780-ee41a9514f05"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.950301 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc8b898-1330-4381-b780-ee41a9514f05-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "efc8b898-1330-4381-b780-ee41a9514f05" (UID: "efc8b898-1330-4381-b780-ee41a9514f05"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:43:00 crc kubenswrapper[4876]: I0313 07:43:00.951245 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efc8b898-1330-4381-b780-ee41a9514f05-kube-api-access-bnljr" (OuterVolumeSpecName: "kube-api-access-bnljr") pod "efc8b898-1330-4381-b780-ee41a9514f05" (UID: "efc8b898-1330-4381-b780-ee41a9514f05"). InnerVolumeSpecName "kube-api-access-bnljr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:01 crc kubenswrapper[4876]: I0313 07:43:01.032984 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efc8b898-1330-4381-b780-ee41a9514f05-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:01 crc kubenswrapper[4876]: I0313 07:43:01.033321 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnljr\" (UniqueName: \"kubernetes.io/projected/efc8b898-1330-4381-b780-ee41a9514f05-kube-api-access-bnljr\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:01 crc kubenswrapper[4876]: I0313 07:43:01.033330 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc8b898-1330-4381-b780-ee41a9514f05-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:01 crc kubenswrapper[4876]: I0313 07:43:01.285163 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:43:01 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:43:01 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:43:01 crc kubenswrapper[4876]: healthz check failed Mar 13 07:43:01 crc kubenswrapper[4876]: I0313 07:43:01.285523 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:43:01 crc kubenswrapper[4876]: I0313 07:43:01.362629 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" event={"ID":"efc8b898-1330-4381-b780-ee41a9514f05","Type":"ContainerDied","Data":"b259609c4cbb383974b53f6f1a6384734c1f9ebf217a9b1f742f3b6812c86700"} Mar 13 07:43:01 crc kubenswrapper[4876]: I0313 07:43:01.362679 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b259609c4cbb383974b53f6f1a6384734c1f9ebf217a9b1f742f3b6812c86700" Mar 13 07:43:01 crc kubenswrapper[4876]: I0313 07:43:01.362791 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.056891 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.056984 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.057024 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.057045 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.059014 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.059202 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.059455 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.069696 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.069768 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.077878 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.092223 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.092753 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.158046 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.158352 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.160372 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.167837 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.174317 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.177208 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f36294fb-98b9-48be-8237-beb6484e0cb6-metrics-certs\") pod \"network-metrics-daemon-zrbh4\" (UID: \"f36294fb-98b9-48be-8237-beb6484e0cb6\") " pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.285415 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:43:02 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:43:02 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:43:02 crc kubenswrapper[4876]: healthz check failed Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.285513 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.415206 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.423532 4876 ???:1] "http: TLS handshake error from 192.168.126.11:54700: no serving certificate available for the kubelet" Mar 13 07:43:02 crc kubenswrapper[4876]: I0313 07:43:02.423623 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zrbh4" Mar 13 07:43:03 crc kubenswrapper[4876]: I0313 07:43:03.290181 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:43:03 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:43:03 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:43:03 crc kubenswrapper[4876]: healthz check failed Mar 13 07:43:03 crc kubenswrapper[4876]: I0313 07:43:03.290619 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:43:03 crc kubenswrapper[4876]: I0313 07:43:03.378288 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zrbh4"] Mar 13 07:43:03 crc kubenswrapper[4876]: I0313 07:43:03.462859 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"58764115c9924711925a1816fd2d37b8aea99029869e786758cfaee6e7476bd7"} Mar 13 07:43:03 crc kubenswrapper[4876]: W0313 07:43:03.571833 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-67ea26085e74f11bb783d724737b96c3f7a292225ac6e526eabed4f6c3d751a2 WatchSource:0}: Error finding container 67ea26085e74f11bb783d724737b96c3f7a292225ac6e526eabed4f6c3d751a2: Status 404 returned error can't find the container with id 67ea26085e74f11bb783d724737b96c3f7a292225ac6e526eabed4f6c3d751a2 Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.302036 4876 patch_prober.go:28] interesting pod/router-default-5444994796-cd4ct container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 07:43:04 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Mar 13 07:43:04 crc kubenswrapper[4876]: [+]process-running ok Mar 13 07:43:04 crc kubenswrapper[4876]: healthz check failed Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.302387 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cd4ct" podUID="1fe6fb32-9851-48de-88f6-94701fcf7010" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.585783 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f7aee021b876e63d3d3db6cfc6d256de8eb42f25f14bca33d9d8f779a4199be5"} Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.585867 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"82e38d1f469413c9e961dc0d59f197631610c60913522bc41835c27f7c944e9e"} Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.586075 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.595610 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" event={"ID":"f36294fb-98b9-48be-8237-beb6484e0cb6","Type":"ContainerStarted","Data":"062387f54f154b445de1930b147d74bb8197443f8432d76c70a28683f01968c4"} Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.595661 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" event={"ID":"f36294fb-98b9-48be-8237-beb6484e0cb6","Type":"ContainerStarted","Data":"68afd394ba65570a23ba80fb39f9c479861c19800fb38d57054befba7887d614"} Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.605284 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d0ee6f159d6c3b0f8c5c7e0b612921a44a4afa2df53b32f16c71db082e935b52"} Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.605387 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"67ea26085e74f11bb783d724737b96c3f7a292225ac6e526eabed4f6c3d751a2"} Mar 13 07:43:04 crc kubenswrapper[4876]: I0313 07:43:04.609423 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d0b6d944be3b06dc6271cdbffc90232902898b8abe946c4485b38ae510be828e"} Mar 13 07:43:05 crc kubenswrapper[4876]: I0313 07:43:05.236679 4876 patch_prober.go:28] interesting pod/console-f9d7485db-g65sf container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Mar 13 07:43:05 crc kubenswrapper[4876]: I0313 07:43:05.237038 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-g65sf" podUID="62fdbab2-2be8-42cc-a936-a87ffd085a0e" containerName="console" probeResult="failure" output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" Mar 13 07:43:05 crc kubenswrapper[4876]: I0313 07:43:05.285538 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:43:05 crc kubenswrapper[4876]: I0313 07:43:05.288494 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-cd4ct" Mar 13 07:43:05 crc kubenswrapper[4876]: I0313 07:43:05.310497 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:05 crc kubenswrapper[4876]: I0313 07:43:05.310553 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:05 crc kubenswrapper[4876]: I0313 07:43:05.310498 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:05 crc kubenswrapper[4876]: I0313 07:43:05.310856 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:12 crc kubenswrapper[4876]: I0313 07:43:12.889181 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-64cd5df9d4-kt45r"] Mar 13 07:43:12 crc kubenswrapper[4876]: I0313 07:43:12.898158 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" podUID="53894f81-37aa-4a7e-837b-0f5e722352af" containerName="controller-manager" containerID="cri-o://417b3d65fa933dbef0d5c6eb42b93643e7d0d0e51d81c6197403aa26835e47e3" gracePeriod=30 Mar 13 07:43:12 crc kubenswrapper[4876]: I0313 07:43:12.901086 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-679d787886-82v44"] Mar 13 07:43:12 crc kubenswrapper[4876]: I0313 07:43:12.901968 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" podUID="b694aa35-ed97-4c15-9161-8deaf40656f6" containerName="route-controller-manager" containerID="cri-o://9fa46fab2dc7890152379da63bb9d40e3ed459ce381562a9736403541ffb4d9b" gracePeriod=30 Mar 13 07:43:13 crc kubenswrapper[4876]: I0313 07:43:13.757805 4876 generic.go:334] "Generic (PLEG): container finished" podID="b694aa35-ed97-4c15-9161-8deaf40656f6" containerID="9fa46fab2dc7890152379da63bb9d40e3ed459ce381562a9736403541ffb4d9b" exitCode=0 Mar 13 07:43:13 crc kubenswrapper[4876]: I0313 07:43:13.757954 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" event={"ID":"b694aa35-ed97-4c15-9161-8deaf40656f6","Type":"ContainerDied","Data":"9fa46fab2dc7890152379da63bb9d40e3ed459ce381562a9736403541ffb4d9b"} Mar 13 07:43:13 crc kubenswrapper[4876]: I0313 07:43:13.761223 4876 generic.go:334] "Generic (PLEG): container finished" podID="53894f81-37aa-4a7e-837b-0f5e722352af" containerID="417b3d65fa933dbef0d5c6eb42b93643e7d0d0e51d81c6197403aa26835e47e3" exitCode=0 Mar 13 07:43:13 crc kubenswrapper[4876]: I0313 07:43:13.761264 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" event={"ID":"53894f81-37aa-4a7e-837b-0f5e722352af","Type":"ContainerDied","Data":"417b3d65fa933dbef0d5c6eb42b93643e7d0d0e51d81c6197403aa26835e47e3"} Mar 13 07:43:14 crc kubenswrapper[4876]: I0313 07:43:14.723452 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.302396 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.302406 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.302443 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.302457 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.302482 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.302957 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"ddea5c6d3e631d9d574fe45141db5454e8f2d38a73b22d3248b2de0c95e65212"} pod="openshift-console/downloads-7954f5f757-6cnvw" containerMessage="Container download-server failed liveness probe, will be restarted" Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.302985 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" containerID="cri-o://ddea5c6d3e631d9d574fe45141db5454e8f2d38a73b22d3248b2de0c95e65212" gracePeriod=2 Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.303034 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.303065 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.318361 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.326327 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.798869 4876 generic.go:334] "Generic (PLEG): container finished" podID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerID="ddea5c6d3e631d9d574fe45141db5454e8f2d38a73b22d3248b2de0c95e65212" exitCode=0 Mar 13 07:43:15 crc kubenswrapper[4876]: I0313 07:43:15.798972 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6cnvw" event={"ID":"ee3187e1-6603-4cd1-a134-6437fae1ca4a","Type":"ContainerDied","Data":"ddea5c6d3e631d9d574fe45141db5454e8f2d38a73b22d3248b2de0c95e65212"} Mar 13 07:43:16 crc kubenswrapper[4876]: I0313 07:43:16.270563 4876 patch_prober.go:28] interesting pod/controller-manager-64cd5df9d4-kt45r container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" start-of-body= Mar 13 07:43:16 crc kubenswrapper[4876]: I0313 07:43:16.270622 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" podUID="53894f81-37aa-4a7e-837b-0f5e722352af" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" Mar 13 07:43:16 crc kubenswrapper[4876]: I0313 07:43:16.493739 4876 patch_prober.go:28] interesting pod/route-controller-manager-679d787886-82v44 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" start-of-body= Mar 13 07:43:16 crc kubenswrapper[4876]: I0313 07:43:16.493791 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" podUID="b694aa35-ed97-4c15-9161-8deaf40656f6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" Mar 13 07:43:21 crc kubenswrapper[4876]: I0313 07:43:21.829910 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zrbh4" event={"ID":"f36294fb-98b9-48be-8237-beb6484e0cb6","Type":"ContainerStarted","Data":"2cad43f69579ea7e0a00c8cae653c393dbb236ed08068101205cf3b9f6747dc5"} Mar 13 07:43:21 crc kubenswrapper[4876]: I0313 07:43:21.846864 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zrbh4" podStartSLOduration=202.846841292 podStartE2EDuration="3m22.846841292s" podCreationTimestamp="2026-03-13 07:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:43:21.841500773 +0000 UTC m=+261.512279755" watchObservedRunningTime="2026-03-13 07:43:21.846841292 +0000 UTC m=+261.517620274" Mar 13 07:43:22 crc kubenswrapper[4876]: E0313 07:43:22.615829 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 13 07:43:22 crc kubenswrapper[4876]: E0313 07:43:22.616211 4876 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 07:43:22 crc kubenswrapper[4876]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 13 07:43:22 crc kubenswrapper[4876]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8wxqx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29556462-xxthz_openshift-infra(01e125dd-87e2-4362-842e-849288698994): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 13 07:43:22 crc kubenswrapper[4876]: > logger="UnhandledError" Mar 13 07:43:22 crc kubenswrapper[4876]: E0313 07:43:22.617396 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29556462-xxthz" podUID="01e125dd-87e2-4362-842e-849288698994" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.626556 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667333 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8"] Mar 13 07:43:22 crc kubenswrapper[4876]: E0313 07:43:22.667624 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc8b898-1330-4381-b780-ee41a9514f05" containerName="collect-profiles" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667643 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc8b898-1330-4381-b780-ee41a9514f05" containerName="collect-profiles" Mar 13 07:43:22 crc kubenswrapper[4876]: E0313 07:43:22.667657 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1578f9c7-03ac-45ac-92d6-cc655178e60f" containerName="pruner" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667664 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1578f9c7-03ac-45ac-92d6-cc655178e60f" containerName="pruner" Mar 13 07:43:22 crc kubenswrapper[4876]: E0313 07:43:22.667672 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d3fa307-6dbc-4387-ab37-903b95f39b93" containerName="pruner" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667679 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d3fa307-6dbc-4387-ab37-903b95f39b93" containerName="pruner" Mar 13 07:43:22 crc kubenswrapper[4876]: E0313 07:43:22.667696 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53894f81-37aa-4a7e-837b-0f5e722352af" containerName="controller-manager" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667703 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="53894f81-37aa-4a7e-837b-0f5e722352af" containerName="controller-manager" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667791 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1578f9c7-03ac-45ac-92d6-cc655178e60f" containerName="pruner" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667807 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d3fa307-6dbc-4387-ab37-903b95f39b93" containerName="pruner" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667817 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="53894f81-37aa-4a7e-837b-0f5e722352af" containerName="controller-manager" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.667829 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="efc8b898-1330-4381-b780-ee41a9514f05" containerName="collect-profiles" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.668219 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.670751 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8"] Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.693682 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-config\") pod \"53894f81-37aa-4a7e-837b-0f5e722352af\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.693764 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53894f81-37aa-4a7e-837b-0f5e722352af-serving-cert\") pod \"53894f81-37aa-4a7e-837b-0f5e722352af\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.693866 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-proxy-ca-bundles\") pod \"53894f81-37aa-4a7e-837b-0f5e722352af\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.693895 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xxhq\" (UniqueName: \"kubernetes.io/projected/53894f81-37aa-4a7e-837b-0f5e722352af-kube-api-access-4xxhq\") pod \"53894f81-37aa-4a7e-837b-0f5e722352af\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.693916 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-client-ca\") pod \"53894f81-37aa-4a7e-837b-0f5e722352af\" (UID: \"53894f81-37aa-4a7e-837b-0f5e722352af\") " Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.694877 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-client-ca" (OuterVolumeSpecName: "client-ca") pod "53894f81-37aa-4a7e-837b-0f5e722352af" (UID: "53894f81-37aa-4a7e-837b-0f5e722352af"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.694919 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-config" (OuterVolumeSpecName: "config") pod "53894f81-37aa-4a7e-837b-0f5e722352af" (UID: "53894f81-37aa-4a7e-837b-0f5e722352af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.694933 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "53894f81-37aa-4a7e-837b-0f5e722352af" (UID: "53894f81-37aa-4a7e-837b-0f5e722352af"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.727363 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53894f81-37aa-4a7e-837b-0f5e722352af-kube-api-access-4xxhq" (OuterVolumeSpecName: "kube-api-access-4xxhq") pod "53894f81-37aa-4a7e-837b-0f5e722352af" (UID: "53894f81-37aa-4a7e-837b-0f5e722352af"). InnerVolumeSpecName "kube-api-access-4xxhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.727365 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53894f81-37aa-4a7e-837b-0f5e722352af-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "53894f81-37aa-4a7e-837b-0f5e722352af" (UID: "53894f81-37aa-4a7e-837b-0f5e722352af"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795037 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f68685f9-4bee-418c-8e82-0cc5c674f4f4-serving-cert\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795118 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4mjq\" (UniqueName: \"kubernetes.io/projected/f68685f9-4bee-418c-8e82-0cc5c674f4f4-kube-api-access-n4mjq\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795151 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-config\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795190 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-client-ca\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795211 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-proxy-ca-bundles\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795262 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53894f81-37aa-4a7e-837b-0f5e722352af-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795321 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795359 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xxhq\" (UniqueName: \"kubernetes.io/projected/53894f81-37aa-4a7e-837b-0f5e722352af-kube-api-access-4xxhq\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795394 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.795412 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53894f81-37aa-4a7e-837b-0f5e722352af-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.838034 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.838097 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64cd5df9d4-kt45r" event={"ID":"53894f81-37aa-4a7e-837b-0f5e722352af","Type":"ContainerDied","Data":"12d09b150a43a7ba5e5e13ced555d4506b54d9ea96d3af93d7ee9f5e2ab7129c"} Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.838164 4876 scope.go:117] "RemoveContainer" containerID="417b3d65fa933dbef0d5c6eb42b93643e7d0d0e51d81c6197403aa26835e47e3" Mar 13 07:43:22 crc kubenswrapper[4876]: E0313 07:43:22.844202 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29556462-xxthz" podUID="01e125dd-87e2-4362-842e-849288698994" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.877154 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-64cd5df9d4-kt45r"] Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.881733 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-64cd5df9d4-kt45r"] Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.897049 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-client-ca\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.897096 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-proxy-ca-bundles\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.897126 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f68685f9-4bee-418c-8e82-0cc5c674f4f4-serving-cert\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.897203 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4mjq\" (UniqueName: \"kubernetes.io/projected/f68685f9-4bee-418c-8e82-0cc5c674f4f4-kube-api-access-n4mjq\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.897273 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-config\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.898250 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-client-ca\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.898865 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-proxy-ca-bundles\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.898929 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-config\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.901070 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f68685f9-4bee-418c-8e82-0cc5c674f4f4-serving-cert\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.916758 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4mjq\" (UniqueName: \"kubernetes.io/projected/f68685f9-4bee-418c-8e82-0cc5c674f4f4-kube-api-access-n4mjq\") pod \"controller-manager-7bcc8d6955-bhgk8\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:22 crc kubenswrapper[4876]: I0313 07:43:22.933951 4876 ???:1] "http: TLS handshake error from 192.168.126.11:57154: no serving certificate available for the kubelet" Mar 13 07:43:23 crc kubenswrapper[4876]: I0313 07:43:23.020699 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:23 crc kubenswrapper[4876]: I0313 07:43:23.044136 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53894f81-37aa-4a7e-837b-0f5e722352af" path="/var/lib/kubelet/pods/53894f81-37aa-4a7e-837b-0f5e722352af/volumes" Mar 13 07:43:24 crc kubenswrapper[4876]: I0313 07:43:24.608298 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:43:24 crc kubenswrapper[4876]: I0313 07:43:24.609446 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:43:25 crc kubenswrapper[4876]: I0313 07:43:25.303299 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:25 crc kubenswrapper[4876]: I0313 07:43:25.303802 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:25 crc kubenswrapper[4876]: I0313 07:43:25.905572 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:43:25 crc kubenswrapper[4876]: I0313 07:43:25.950346 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt"] Mar 13 07:43:25 crc kubenswrapper[4876]: E0313 07:43:25.951007 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b694aa35-ed97-4c15-9161-8deaf40656f6" containerName="route-controller-manager" Mar 13 07:43:25 crc kubenswrapper[4876]: I0313 07:43:25.951028 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b694aa35-ed97-4c15-9161-8deaf40656f6" containerName="route-controller-manager" Mar 13 07:43:25 crc kubenswrapper[4876]: I0313 07:43:25.951174 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b694aa35-ed97-4c15-9161-8deaf40656f6" containerName="route-controller-manager" Mar 13 07:43:25 crc kubenswrapper[4876]: I0313 07:43:25.951612 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt"] Mar 13 07:43:25 crc kubenswrapper[4876]: I0313 07:43:25.951704 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.040517 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b694aa35-ed97-4c15-9161-8deaf40656f6-serving-cert\") pod \"b694aa35-ed97-4c15-9161-8deaf40656f6\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.040573 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hqlf\" (UniqueName: \"kubernetes.io/projected/b694aa35-ed97-4c15-9161-8deaf40656f6-kube-api-access-5hqlf\") pod \"b694aa35-ed97-4c15-9161-8deaf40656f6\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.040690 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-client-ca\") pod \"b694aa35-ed97-4c15-9161-8deaf40656f6\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.040829 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-config\") pod \"b694aa35-ed97-4c15-9161-8deaf40656f6\" (UID: \"b694aa35-ed97-4c15-9161-8deaf40656f6\") " Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.041020 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-client-ca\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.041055 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4999caff-c27b-4106-ab4b-33d0fcd37530-serving-cert\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.041134 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-config\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.041178 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmtfk\" (UniqueName: \"kubernetes.io/projected/4999caff-c27b-4106-ab4b-33d0fcd37530-kube-api-access-vmtfk\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.041986 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-config" (OuterVolumeSpecName: "config") pod "b694aa35-ed97-4c15-9161-8deaf40656f6" (UID: "b694aa35-ed97-4c15-9161-8deaf40656f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.042176 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-client-ca" (OuterVolumeSpecName: "client-ca") pod "b694aa35-ed97-4c15-9161-8deaf40656f6" (UID: "b694aa35-ed97-4c15-9161-8deaf40656f6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.047311 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b694aa35-ed97-4c15-9161-8deaf40656f6-kube-api-access-5hqlf" (OuterVolumeSpecName: "kube-api-access-5hqlf") pod "b694aa35-ed97-4c15-9161-8deaf40656f6" (UID: "b694aa35-ed97-4c15-9161-8deaf40656f6"). InnerVolumeSpecName "kube-api-access-5hqlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.047801 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b694aa35-ed97-4c15-9161-8deaf40656f6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b694aa35-ed97-4c15-9161-8deaf40656f6" (UID: "b694aa35-ed97-4c15-9161-8deaf40656f6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.132479 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wbv9l" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.142119 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-config\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.143690 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmtfk\" (UniqueName: \"kubernetes.io/projected/4999caff-c27b-4106-ab4b-33d0fcd37530-kube-api-access-vmtfk\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.143867 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-client-ca\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.144039 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4999caff-c27b-4106-ab4b-33d0fcd37530-serving-cert\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.144638 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-client-ca\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.144706 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.145444 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b694aa35-ed97-4c15-9161-8deaf40656f6-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.145469 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b694aa35-ed97-4c15-9161-8deaf40656f6-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.145491 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hqlf\" (UniqueName: \"kubernetes.io/projected/b694aa35-ed97-4c15-9161-8deaf40656f6-kube-api-access-5hqlf\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.148158 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-config\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.156025 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4999caff-c27b-4106-ab4b-33d0fcd37530-serving-cert\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.160795 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmtfk\" (UniqueName: \"kubernetes.io/projected/4999caff-c27b-4106-ab4b-33d0fcd37530-kube-api-access-vmtfk\") pod \"route-controller-manager-79bffd544d-xd9vt\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.270347 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.866692 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" event={"ID":"b694aa35-ed97-4c15-9161-8deaf40656f6","Type":"ContainerDied","Data":"4d446e976750c4d5bd1558b023f74ee2b6b61ab2f9cbe6c312341b1606b5bcc4"} Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.866780 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-679d787886-82v44" Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.898234 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-679d787886-82v44"] Mar 13 07:43:26 crc kubenswrapper[4876]: I0313 07:43:26.905774 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-679d787886-82v44"] Mar 13 07:43:27 crc kubenswrapper[4876]: I0313 07:43:27.042132 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b694aa35-ed97-4c15-9161-8deaf40656f6" path="/var/lib/kubelet/pods/b694aa35-ed97-4c15-9161-8deaf40656f6/volumes" Mar 13 07:43:28 crc kubenswrapper[4876]: E0313 07:43:28.297587 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 07:43:28 crc kubenswrapper[4876]: E0313 07:43:28.298174 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zrvnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-mtcn4_openshift-marketplace(18c1ffa5-3580-48f1-8df2-5f442d47062d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 07:43:28 crc kubenswrapper[4876]: E0313 07:43:28.299794 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-mtcn4" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.662079 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.662971 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.665620 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.665763 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.677721 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.817567 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfb8a516-01c2-4801-aede-9176159ec93e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dfb8a516-01c2-4801-aede-9176159ec93e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.817641 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfb8a516-01c2-4801-aede-9176159ec93e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dfb8a516-01c2-4801-aede-9176159ec93e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.918414 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfb8a516-01c2-4801-aede-9176159ec93e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dfb8a516-01c2-4801-aede-9176159ec93e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.918511 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfb8a516-01c2-4801-aede-9176159ec93e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dfb8a516-01c2-4801-aede-9176159ec93e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.918558 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfb8a516-01c2-4801-aede-9176159ec93e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dfb8a516-01c2-4801-aede-9176159ec93e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:28 crc kubenswrapper[4876]: I0313 07:43:28.949064 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfb8a516-01c2-4801-aede-9176159ec93e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dfb8a516-01c2-4801-aede-9176159ec93e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:29 crc kubenswrapper[4876]: I0313 07:43:29.035149 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:30 crc kubenswrapper[4876]: E0313 07:43:30.144170 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-mtcn4" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" Mar 13 07:43:30 crc kubenswrapper[4876]: E0313 07:43:30.221605 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 07:43:30 crc kubenswrapper[4876]: E0313 07:43:30.222048 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zl8wt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kgzdt_openshift-marketplace(af6d5722-1354-4021-9ba4-6ad9a1286c60): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 07:43:30 crc kubenswrapper[4876]: E0313 07:43:30.225850 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kgzdt" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.859676 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.860402 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.868189 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-var-lock\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.868282 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.868307 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de3aea18-c726-431d-97ed-af1a5341ffb5-kube-api-access\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.873353 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.892153 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8"] Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.969527 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.969579 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de3aea18-c726-431d-97ed-af1a5341ffb5-kube-api-access\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.969649 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-var-lock\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.969704 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-var-lock\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.969701 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.978134 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt"] Mar 13 07:43:32 crc kubenswrapper[4876]: I0313 07:43:32.998835 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de3aea18-c726-431d-97ed-af1a5341ffb5-kube-api-access\") pod \"installer-9-crc\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:33 crc kubenswrapper[4876]: I0313 07:43:33.220798 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:43:35 crc kubenswrapper[4876]: E0313 07:43:35.198616 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kgzdt" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" Mar 13 07:43:35 crc kubenswrapper[4876]: I0313 07:43:35.303339 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:35 crc kubenswrapper[4876]: I0313 07:43:35.303842 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:35 crc kubenswrapper[4876]: E0313 07:43:35.308680 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 07:43:35 crc kubenswrapper[4876]: E0313 07:43:35.308858 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v7bxl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gkxcc_openshift-marketplace(68fd169c-1d47-4f5b-a528-201a1ea159a1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 07:43:35 crc kubenswrapper[4876]: E0313 07:43:35.311391 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gkxcc" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" Mar 13 07:43:36 crc kubenswrapper[4876]: E0313 07:43:36.798172 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gkxcc" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" Mar 13 07:43:36 crc kubenswrapper[4876]: I0313 07:43:36.805789 4876 scope.go:117] "RemoveContainer" containerID="9fa46fab2dc7890152379da63bb9d40e3ed459ce381562a9736403541ffb4d9b" Mar 13 07:43:36 crc kubenswrapper[4876]: E0313 07:43:36.988720 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 07:43:36 crc kubenswrapper[4876]: E0313 07:43:36.989209 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r5lkh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bg8rq_openshift-marketplace(c4259aab-8c71-4b40-957e-ef4e5ddc1ea1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 07:43:36 crc kubenswrapper[4876]: E0313 07:43:36.990851 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bg8rq" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.107569 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.107726 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rr79z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cf5vz_openshift-marketplace(ddbd46d5-05cd-4d71-b085-142b13687f63): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.110031 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cf5vz" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.121613 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt"] Mar 13 07:43:37 crc kubenswrapper[4876]: W0313 07:43:37.230108 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4999caff_c27b_4106_ab4b_33d0fcd37530.slice/crio-291c385f0a65b5a159acb1ced1cc498e86d53a912e8c62c9f1640cdfba698755 WatchSource:0}: Error finding container 291c385f0a65b5a159acb1ced1cc498e86d53a912e8c62c9f1640cdfba698755: Status 404 returned error can't find the container with id 291c385f0a65b5a159acb1ced1cc498e86d53a912e8c62c9f1640cdfba698755 Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.276856 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.277008 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h95dx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kdxtz_openshift-marketplace(f7fbdc82-2033-4493-8f3a-ee17b88eb94e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.278265 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kdxtz" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.290078 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.403316 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.412350 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8"] Mar 13 07:43:37 crc kubenswrapper[4876]: W0313 07:43:37.435169 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf68685f9_4bee_418c_8e82_0cc5c674f4f4.slice/crio-45b747bcd57697d7e1534a5eb520835ea280c1a75f7ad5df2ea6840531a15583 WatchSource:0}: Error finding container 45b747bcd57697d7e1534a5eb520835ea280c1a75f7ad5df2ea6840531a15583: Status 404 returned error can't find the container with id 45b747bcd57697d7e1534a5eb520835ea280c1a75f7ad5df2ea6840531a15583 Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.632048 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.632490 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ws8qb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2q97b_openshift-marketplace(63f3a5db-0db1-4486-a982-39c60bda9dc4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.634504 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2q97b" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.956872 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" event={"ID":"f68685f9-4bee-418c-8e82-0cc5c674f4f4","Type":"ContainerStarted","Data":"3fd17f9a4ded5706d5a26792a484912481f46bc41b5c9854d5f61914d6af7696"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.958190 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.957274 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" podUID="f68685f9-4bee-418c-8e82-0cc5c674f4f4" containerName="controller-manager" containerID="cri-o://3fd17f9a4ded5706d5a26792a484912481f46bc41b5c9854d5f61914d6af7696" gracePeriod=30 Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.958306 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" event={"ID":"f68685f9-4bee-418c-8e82-0cc5c674f4f4","Type":"ContainerStarted","Data":"45b747bcd57697d7e1534a5eb520835ea280c1a75f7ad5df2ea6840531a15583"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.960025 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"de3aea18-c726-431d-97ed-af1a5341ffb5","Type":"ContainerStarted","Data":"87f4ab1a9aaecfbeb6146e65bbd991b5eb6b25ee11b77b665e265cfaf6ea4cfe"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.960073 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"de3aea18-c726-431d-97ed-af1a5341ffb5","Type":"ContainerStarted","Data":"832be911592e1c05faf0c48dc92f243f41b4a87d90e21a017b640f2fa0202a3f"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.960196 4876 patch_prober.go:28] interesting pod/controller-manager-7bcc8d6955-bhgk8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.960264 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" podUID="f68685f9-4bee-418c-8e82-0cc5c674f4f4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.966971 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6cnvw" event={"ID":"ee3187e1-6603-4cd1-a134-6437fae1ca4a","Type":"ContainerStarted","Data":"a1421cee2f7d81e4f9bdf325ed370b536cd6de094679909368804aa56f88dc68"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.967468 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.967901 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.968051 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.968719 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" event={"ID":"4999caff-c27b-4106-ab4b-33d0fcd37530","Type":"ContainerStarted","Data":"c2733122b5ab230ed9236560bd31b1c3607fc1e8a98d7fd616654db56605717b"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.968820 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" event={"ID":"4999caff-c27b-4106-ab4b-33d0fcd37530","Type":"ContainerStarted","Data":"291c385f0a65b5a159acb1ced1cc498e86d53a912e8c62c9f1640cdfba698755"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.968936 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" podUID="4999caff-c27b-4106-ab4b-33d0fcd37530" containerName="route-controller-manager" containerID="cri-o://c2733122b5ab230ed9236560bd31b1c3607fc1e8a98d7fd616654db56605717b" gracePeriod=30 Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.969152 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.973382 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dfb8a516-01c2-4801-aede-9176159ec93e","Type":"ContainerStarted","Data":"7de9b15bdc07f383631672fb50d65e56e28375f505dcbe0d80a710a23bb10ed2"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.973434 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dfb8a516-01c2-4801-aede-9176159ec93e","Type":"ContainerStarted","Data":"183d2c8524d38690e7846a9d375f9852db9cc8dfcab4e9794e13c6c905b099c4"} Mar 13 07:43:37 crc kubenswrapper[4876]: I0313 07:43:37.978652 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556462-xxthz" event={"ID":"01e125dd-87e2-4362-842e-849288698994","Type":"ContainerStarted","Data":"5a8c1853fd57f8e91f5ac9a13a6a8e9950f23601731d0d4b9947ae2713569eed"} Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.980502 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cf5vz" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.981019 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kdxtz" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.981051 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2q97b" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" Mar 13 07:43:37 crc kubenswrapper[4876]: E0313 07:43:37.981028 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bg8rq" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.002917 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" podStartSLOduration=26.002895591 podStartE2EDuration="26.002895591s" podCreationTimestamp="2026-03-13 07:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:43:37.983830949 +0000 UTC m=+277.654609931" watchObservedRunningTime="2026-03-13 07:43:38.002895591 +0000 UTC m=+277.673674573" Mar 13 07:43:38 crc kubenswrapper[4876]: E0313 07:43:38.091013 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 07:43:38 crc kubenswrapper[4876]: E0313 07:43:38.091775 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9ld7c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-fmcl5_openshift-marketplace(8c486d8f-87cf-42c0-9b8c-8f715755edc8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 07:43:38 crc kubenswrapper[4876]: E0313 07:43:38.093122 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-fmcl5" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.104001 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" podStartSLOduration=26.10397326 podStartE2EDuration="26.10397326s" podCreationTimestamp="2026-03-13 07:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:43:38.101652523 +0000 UTC m=+277.772431505" watchObservedRunningTime="2026-03-13 07:43:38.10397326 +0000 UTC m=+277.774752242" Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.118398 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556462-xxthz" podStartSLOduration=47.443255511 podStartE2EDuration="1m38.118381157s" podCreationTimestamp="2026-03-13 07:42:00 +0000 UTC" firstStartedPulling="2026-03-13 07:42:46.575497511 +0000 UTC m=+226.246276493" lastFinishedPulling="2026-03-13 07:43:37.250623157 +0000 UTC m=+276.921402139" observedRunningTime="2026-03-13 07:43:38.1169767 +0000 UTC m=+277.787755682" watchObservedRunningTime="2026-03-13 07:43:38.118381157 +0000 UTC m=+277.789160139" Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.262293 4876 patch_prober.go:28] interesting pod/route-controller-manager-79bffd544d-xd9vt container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": read tcp 10.217.0.2:46864->10.217.0.58:8443: read: connection reset by peer" start-of-body= Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.262392 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" podUID="4999caff-c27b-4106-ab4b-33d0fcd37530" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": read tcp 10.217.0.2:46864->10.217.0.58:8443: read: connection reset by peer" Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.375563 4876 csr.go:261] certificate signing request csr-8vhl6 is approved, waiting to be issued Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.384046 4876 csr.go:257] certificate signing request csr-8vhl6 is issued Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.983774 4876 generic.go:334] "Generic (PLEG): container finished" podID="f68685f9-4bee-418c-8e82-0cc5c674f4f4" containerID="3fd17f9a4ded5706d5a26792a484912481f46bc41b5c9854d5f61914d6af7696" exitCode=0 Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.983830 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" event={"ID":"f68685f9-4bee-418c-8e82-0cc5c674f4f4","Type":"ContainerDied","Data":"3fd17f9a4ded5706d5a26792a484912481f46bc41b5c9854d5f61914d6af7696"} Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.983858 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" event={"ID":"f68685f9-4bee-418c-8e82-0cc5c674f4f4","Type":"ContainerDied","Data":"45b747bcd57697d7e1534a5eb520835ea280c1a75f7ad5df2ea6840531a15583"} Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.983868 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45b747bcd57697d7e1534a5eb520835ea280c1a75f7ad5df2ea6840531a15583" Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.990431 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-79bffd544d-xd9vt_4999caff-c27b-4106-ab4b-33d0fcd37530/route-controller-manager/0.log" Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.990892 4876 generic.go:334] "Generic (PLEG): container finished" podID="4999caff-c27b-4106-ab4b-33d0fcd37530" containerID="c2733122b5ab230ed9236560bd31b1c3607fc1e8a98d7fd616654db56605717b" exitCode=255 Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.991110 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" event={"ID":"4999caff-c27b-4106-ab4b-33d0fcd37530","Type":"ContainerDied","Data":"c2733122b5ab230ed9236560bd31b1c3607fc1e8a98d7fd616654db56605717b"} Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.992788 4876 generic.go:334] "Generic (PLEG): container finished" podID="dfb8a516-01c2-4801-aede-9176159ec93e" containerID="7de9b15bdc07f383631672fb50d65e56e28375f505dcbe0d80a710a23bb10ed2" exitCode=0 Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.992845 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dfb8a516-01c2-4801-aede-9176159ec93e","Type":"ContainerDied","Data":"7de9b15bdc07f383631672fb50d65e56e28375f505dcbe0d80a710a23bb10ed2"} Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.997923 4876 generic.go:334] "Generic (PLEG): container finished" podID="01e125dd-87e2-4362-842e-849288698994" containerID="5a8c1853fd57f8e91f5ac9a13a6a8e9950f23601731d0d4b9947ae2713569eed" exitCode=0 Mar 13 07:43:38 crc kubenswrapper[4876]: I0313 07:43:38.998313 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556462-xxthz" event={"ID":"01e125dd-87e2-4362-842e-849288698994","Type":"ContainerDied","Data":"5a8c1853fd57f8e91f5ac9a13a6a8e9950f23601731d0d4b9947ae2713569eed"} Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:38.999294 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:38.999336 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:39 crc kubenswrapper[4876]: E0313 07:43:39.000481 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-fmcl5" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.035519 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=7.035503792 podStartE2EDuration="7.035503792s" podCreationTimestamp="2026-03-13 07:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:43:39.034708034 +0000 UTC m=+278.705487016" watchObservedRunningTime="2026-03-13 07:43:39.035503792 +0000 UTC m=+278.706282774" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.056160 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.132388 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f9775c99c-6vjh2"] Mar 13 07:43:39 crc kubenswrapper[4876]: E0313 07:43:39.132667 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f68685f9-4bee-418c-8e82-0cc5c674f4f4" containerName="controller-manager" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.132680 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f68685f9-4bee-418c-8e82-0cc5c674f4f4" containerName="controller-manager" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.132777 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f68685f9-4bee-418c-8e82-0cc5c674f4f4" containerName="controller-manager" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.133310 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.140941 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f9775c99c-6vjh2"] Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.180504 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-client-ca\") pod \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.180553 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f68685f9-4bee-418c-8e82-0cc5c674f4f4-serving-cert\") pod \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.180584 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-config\") pod \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.180609 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4mjq\" (UniqueName: \"kubernetes.io/projected/f68685f9-4bee-418c-8e82-0cc5c674f4f4-kube-api-access-n4mjq\") pod \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.180694 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-proxy-ca-bundles\") pod \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\" (UID: \"f68685f9-4bee-418c-8e82-0cc5c674f4f4\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.182668 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-config\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.182800 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsrfz\" (UniqueName: \"kubernetes.io/projected/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-kube-api-access-wsrfz\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.182834 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-client-ca\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.182831 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f68685f9-4bee-418c-8e82-0cc5c674f4f4" (UID: "f68685f9-4bee-418c-8e82-0cc5c674f4f4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.182889 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-proxy-ca-bundles\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.182922 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-serving-cert\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.182964 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.183569 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-config" (OuterVolumeSpecName: "config") pod "f68685f9-4bee-418c-8e82-0cc5c674f4f4" (UID: "f68685f9-4bee-418c-8e82-0cc5c674f4f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.185104 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-client-ca" (OuterVolumeSpecName: "client-ca") pod "f68685f9-4bee-418c-8e82-0cc5c674f4f4" (UID: "f68685f9-4bee-418c-8e82-0cc5c674f4f4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.186940 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f68685f9-4bee-418c-8e82-0cc5c674f4f4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f68685f9-4bee-418c-8e82-0cc5c674f4f4" (UID: "f68685f9-4bee-418c-8e82-0cc5c674f4f4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.187670 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f68685f9-4bee-418c-8e82-0cc5c674f4f4-kube-api-access-n4mjq" (OuterVolumeSpecName: "kube-api-access-n4mjq") pod "f68685f9-4bee-418c-8e82-0cc5c674f4f4" (UID: "f68685f9-4bee-418c-8e82-0cc5c674f4f4"). InnerVolumeSpecName "kube-api-access-n4mjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.223057 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-79bffd544d-xd9vt_4999caff-c27b-4106-ab4b-33d0fcd37530/route-controller-manager/0.log" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.223367 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.283870 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-config\") pod \"4999caff-c27b-4106-ab4b-33d0fcd37530\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.284360 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmtfk\" (UniqueName: \"kubernetes.io/projected/4999caff-c27b-4106-ab4b-33d0fcd37530-kube-api-access-vmtfk\") pod \"4999caff-c27b-4106-ab4b-33d0fcd37530\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.284488 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-client-ca\") pod \"4999caff-c27b-4106-ab4b-33d0fcd37530\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.284961 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-client-ca" (OuterVolumeSpecName: "client-ca") pod "4999caff-c27b-4106-ab4b-33d0fcd37530" (UID: "4999caff-c27b-4106-ab4b-33d0fcd37530"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.285004 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-config" (OuterVolumeSpecName: "config") pod "4999caff-c27b-4106-ab4b-33d0fcd37530" (UID: "4999caff-c27b-4106-ab4b-33d0fcd37530"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.285039 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4999caff-c27b-4106-ab4b-33d0fcd37530-serving-cert\") pod \"4999caff-c27b-4106-ab4b-33d0fcd37530\" (UID: \"4999caff-c27b-4106-ab4b-33d0fcd37530\") " Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286089 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-serving-cert\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286141 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-config\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286221 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsrfz\" (UniqueName: \"kubernetes.io/projected/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-kube-api-access-wsrfz\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286273 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-client-ca\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286329 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-proxy-ca-bundles\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286373 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286383 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f68685f9-4bee-418c-8e82-0cc5c674f4f4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286393 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286403 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4mjq\" (UniqueName: \"kubernetes.io/projected/f68685f9-4bee-418c-8e82-0cc5c674f4f4-kube-api-access-n4mjq\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286414 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4999caff-c27b-4106-ab4b-33d0fcd37530-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.286427 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f68685f9-4bee-418c-8e82-0cc5c674f4f4-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.287879 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-proxy-ca-bundles\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.292316 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-client-ca\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.296375 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-config\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.296440 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4999caff-c27b-4106-ab4b-33d0fcd37530-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4999caff-c27b-4106-ab4b-33d0fcd37530" (UID: "4999caff-c27b-4106-ab4b-33d0fcd37530"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.296636 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4999caff-c27b-4106-ab4b-33d0fcd37530-kube-api-access-vmtfk" (OuterVolumeSpecName: "kube-api-access-vmtfk") pod "4999caff-c27b-4106-ab4b-33d0fcd37530" (UID: "4999caff-c27b-4106-ab4b-33d0fcd37530"). InnerVolumeSpecName "kube-api-access-vmtfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.297415 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-serving-cert\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.305960 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsrfz\" (UniqueName: \"kubernetes.io/projected/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-kube-api-access-wsrfz\") pod \"controller-manager-5f9775c99c-6vjh2\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.385368 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-09 22:50:46.058768321 +0000 UTC Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.385418 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6519h7m6.673354377s for next certificate rotation Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.387961 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmtfk\" (UniqueName: \"kubernetes.io/projected/4999caff-c27b-4106-ab4b-33d0fcd37530-kube-api-access-vmtfk\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.388006 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4999caff-c27b-4106-ab4b-33d0fcd37530-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.464571 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:39 crc kubenswrapper[4876]: I0313 07:43:39.658624 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f9775c99c-6vjh2"] Mar 13 07:43:39 crc kubenswrapper[4876]: W0313 07:43:39.670138 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd32a8720_cb92_4f9b_8cbb_7feeffc07ebf.slice/crio-bbc51cc8c7a2c8a73973bd5e5c3b53719735da18fada393e13434779a4bf1918 WatchSource:0}: Error finding container bbc51cc8c7a2c8a73973bd5e5c3b53719735da18fada393e13434779a4bf1918: Status 404 returned error can't find the container with id bbc51cc8c7a2c8a73973bd5e5c3b53719735da18fada393e13434779a4bf1918 Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.005467 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-79bffd544d-xd9vt_4999caff-c27b-4106-ab4b-33d0fcd37530/route-controller-manager/0.log" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.005581 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" event={"ID":"4999caff-c27b-4106-ab4b-33d0fcd37530","Type":"ContainerDied","Data":"291c385f0a65b5a159acb1ced1cc498e86d53a912e8c62c9f1640cdfba698755"} Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.005610 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.005867 4876 scope.go:117] "RemoveContainer" containerID="c2733122b5ab230ed9236560bd31b1c3607fc1e8a98d7fd616654db56605717b" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.012092 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" event={"ID":"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf","Type":"ContainerStarted","Data":"7a6d79b5c29ad20927dfc833471e1600538e5c984105a902449bed11f065232b"} Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.012163 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" event={"ID":"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf","Type":"ContainerStarted","Data":"bbc51cc8c7a2c8a73973bd5e5c3b53719735da18fada393e13434779a4bf1918"} Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.012279 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.013409 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.034123 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.064879 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" podStartSLOduration=8.064861382 podStartE2EDuration="8.064861382s" podCreationTimestamp="2026-03-13 07:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:43:40.054330746 +0000 UTC m=+279.725109738" watchObservedRunningTime="2026-03-13 07:43:40.064861382 +0000 UTC m=+279.735640364" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.068935 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8"] Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.072485 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7bcc8d6955-bhgk8"] Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.103072 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt"] Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.115508 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79bffd544d-xd9vt"] Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.337531 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556462-xxthz" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.386761 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-14 12:41:14.256847183 +0000 UTC Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.386804 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7372h57m33.870045652s for next certificate rotation Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.431472 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.499981 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wxqx\" (UniqueName: \"kubernetes.io/projected/01e125dd-87e2-4362-842e-849288698994-kube-api-access-8wxqx\") pod \"01e125dd-87e2-4362-842e-849288698994\" (UID: \"01e125dd-87e2-4362-842e-849288698994\") " Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.505679 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e125dd-87e2-4362-842e-849288698994-kube-api-access-8wxqx" (OuterVolumeSpecName: "kube-api-access-8wxqx") pod "01e125dd-87e2-4362-842e-849288698994" (UID: "01e125dd-87e2-4362-842e-849288698994"). InnerVolumeSpecName "kube-api-access-8wxqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.601093 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfb8a516-01c2-4801-aede-9176159ec93e-kubelet-dir\") pod \"dfb8a516-01c2-4801-aede-9176159ec93e\" (UID: \"dfb8a516-01c2-4801-aede-9176159ec93e\") " Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.601256 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfb8a516-01c2-4801-aede-9176159ec93e-kube-api-access\") pod \"dfb8a516-01c2-4801-aede-9176159ec93e\" (UID: \"dfb8a516-01c2-4801-aede-9176159ec93e\") " Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.601235 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dfb8a516-01c2-4801-aede-9176159ec93e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dfb8a516-01c2-4801-aede-9176159ec93e" (UID: "dfb8a516-01c2-4801-aede-9176159ec93e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.601556 4876 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfb8a516-01c2-4801-aede-9176159ec93e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.601570 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wxqx\" (UniqueName: \"kubernetes.io/projected/01e125dd-87e2-4362-842e-849288698994-kube-api-access-8wxqx\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.604909 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb8a516-01c2-4801-aede-9176159ec93e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dfb8a516-01c2-4801-aede-9176159ec93e" (UID: "dfb8a516-01c2-4801-aede-9176159ec93e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:40 crc kubenswrapper[4876]: I0313 07:43:40.702687 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfb8a516-01c2-4801-aede-9176159ec93e-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.019698 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dfb8a516-01c2-4801-aede-9176159ec93e","Type":"ContainerDied","Data":"183d2c8524d38690e7846a9d375f9852db9cc8dfcab4e9794e13c6c905b099c4"} Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.019735 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.019742 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="183d2c8524d38690e7846a9d375f9852db9cc8dfcab4e9794e13c6c905b099c4" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.021817 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556462-xxthz" event={"ID":"01e125dd-87e2-4362-842e-849288698994","Type":"ContainerDied","Data":"005a74e6f6269bd9bf747ff7c6885857c52d9bda3e2efb70a162e8e5b4e88f6c"} Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.021860 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="005a74e6f6269bd9bf747ff7c6885857c52d9bda3e2efb70a162e8e5b4e88f6c" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.021877 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556462-xxthz" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.043216 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4999caff-c27b-4106-ab4b-33d0fcd37530" path="/var/lib/kubelet/pods/4999caff-c27b-4106-ab4b-33d0fcd37530/volumes" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.044903 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f68685f9-4bee-418c-8e82-0cc5c674f4f4" path="/var/lib/kubelet/pods/f68685f9-4bee-418c-8e82-0cc5c674f4f4/volumes" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.955699 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6"] Mar 13 07:43:41 crc kubenswrapper[4876]: E0313 07:43:41.955937 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb8a516-01c2-4801-aede-9176159ec93e" containerName="pruner" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.955955 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb8a516-01c2-4801-aede-9176159ec93e" containerName="pruner" Mar 13 07:43:41 crc kubenswrapper[4876]: E0313 07:43:41.955969 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e125dd-87e2-4362-842e-849288698994" containerName="oc" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.955978 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e125dd-87e2-4362-842e-849288698994" containerName="oc" Mar 13 07:43:41 crc kubenswrapper[4876]: E0313 07:43:41.955992 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4999caff-c27b-4106-ab4b-33d0fcd37530" containerName="route-controller-manager" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.955999 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4999caff-c27b-4106-ab4b-33d0fcd37530" containerName="route-controller-manager" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.956119 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e125dd-87e2-4362-842e-849288698994" containerName="oc" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.956135 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4999caff-c27b-4106-ab4b-33d0fcd37530" containerName="route-controller-manager" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.956141 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb8a516-01c2-4801-aede-9176159ec93e" containerName="pruner" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.956529 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.958509 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.958551 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.958835 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.958957 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.958980 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.959956 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 07:43:41 crc kubenswrapper[4876]: I0313 07:43:41.967881 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6"] Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.121593 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-client-ca\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.121686 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fftk2\" (UniqueName: \"kubernetes.io/projected/60d8edcc-191c-4651-818f-54b6a324837f-kube-api-access-fftk2\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.121983 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d8edcc-191c-4651-818f-54b6a324837f-serving-cert\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.122337 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-config\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.178854 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.224761 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-client-ca\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.224870 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fftk2\" (UniqueName: \"kubernetes.io/projected/60d8edcc-191c-4651-818f-54b6a324837f-kube-api-access-fftk2\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.224905 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d8edcc-191c-4651-818f-54b6a324837f-serving-cert\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.225337 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-config\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.225701 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-client-ca\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.226554 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-config\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.238946 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d8edcc-191c-4651-818f-54b6a324837f-serving-cert\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.248221 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fftk2\" (UniqueName: \"kubernetes.io/projected/60d8edcc-191c-4651-818f-54b6a324837f-kube-api-access-fftk2\") pod \"route-controller-manager-756688789b-z9ml6\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.272960 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:42 crc kubenswrapper[4876]: I0313 07:43:42.778774 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6"] Mar 13 07:43:42 crc kubenswrapper[4876]: W0313 07:43:42.810744 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60d8edcc_191c_4651_818f_54b6a324837f.slice/crio-1d8093014768ff7ae73a183e0f41c47551bad9d9b665f4e9eb5ccad826f5cecd WatchSource:0}: Error finding container 1d8093014768ff7ae73a183e0f41c47551bad9d9b665f4e9eb5ccad826f5cecd: Status 404 returned error can't find the container with id 1d8093014768ff7ae73a183e0f41c47551bad9d9b665f4e9eb5ccad826f5cecd Mar 13 07:43:43 crc kubenswrapper[4876]: I0313 07:43:43.049000 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" event={"ID":"60d8edcc-191c-4651-818f-54b6a324837f","Type":"ContainerStarted","Data":"6232c2487036bf7a8f7aa71d5f759be41146432089422b0210201f88353f54a5"} Mar 13 07:43:43 crc kubenswrapper[4876]: I0313 07:43:43.049035 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" event={"ID":"60d8edcc-191c-4651-818f-54b6a324837f","Type":"ContainerStarted","Data":"1d8093014768ff7ae73a183e0f41c47551bad9d9b665f4e9eb5ccad826f5cecd"} Mar 13 07:43:43 crc kubenswrapper[4876]: I0313 07:43:43.050018 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:43 crc kubenswrapper[4876]: I0313 07:43:43.050618 4876 patch_prober.go:28] interesting pod/route-controller-manager-756688789b-z9ml6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.63:8443/healthz\": dial tcp 10.217.0.63:8443: connect: connection refused" start-of-body= Mar 13 07:43:43 crc kubenswrapper[4876]: I0313 07:43:43.050658 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" podUID="60d8edcc-191c-4651-818f-54b6a324837f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.63:8443/healthz\": dial tcp 10.217.0.63:8443: connect: connection refused" Mar 13 07:43:43 crc kubenswrapper[4876]: I0313 07:43:43.052852 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtcn4" event={"ID":"18c1ffa5-3580-48f1-8df2-5f442d47062d","Type":"ContainerStarted","Data":"8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7"} Mar 13 07:43:43 crc kubenswrapper[4876]: I0313 07:43:43.075103 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" podStartSLOduration=10.075085368 podStartE2EDuration="10.075085368s" podCreationTimestamp="2026-03-13 07:43:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:43:43.074940353 +0000 UTC m=+282.745719335" watchObservedRunningTime="2026-03-13 07:43:43.075085368 +0000 UTC m=+282.745864350" Mar 13 07:43:44 crc kubenswrapper[4876]: I0313 07:43:44.062291 4876 generic.go:334] "Generic (PLEG): container finished" podID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerID="8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7" exitCode=0 Mar 13 07:43:44 crc kubenswrapper[4876]: I0313 07:43:44.062382 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtcn4" event={"ID":"18c1ffa5-3580-48f1-8df2-5f442d47062d","Type":"ContainerDied","Data":"8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7"} Mar 13 07:43:44 crc kubenswrapper[4876]: I0313 07:43:44.069811 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:45 crc kubenswrapper[4876]: I0313 07:43:45.301920 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:45 crc kubenswrapper[4876]: I0313 07:43:45.301918 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-6cnvw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 13 07:43:45 crc kubenswrapper[4876]: I0313 07:43:45.301977 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:45 crc kubenswrapper[4876]: I0313 07:43:45.302026 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6cnvw" podUID="ee3187e1-6603-4cd1-a134-6437fae1ca4a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 13 07:43:46 crc kubenswrapper[4876]: I0313 07:43:46.074826 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtcn4" event={"ID":"18c1ffa5-3580-48f1-8df2-5f442d47062d","Type":"ContainerStarted","Data":"9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b"} Mar 13 07:43:46 crc kubenswrapper[4876]: I0313 07:43:46.094940 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mtcn4" podStartSLOduration=3.330211217 podStartE2EDuration="53.094922348s" podCreationTimestamp="2026-03-13 07:42:53 +0000 UTC" firstStartedPulling="2026-03-13 07:42:55.867738547 +0000 UTC m=+235.538517539" lastFinishedPulling="2026-03-13 07:43:45.632449688 +0000 UTC m=+285.303228670" observedRunningTime="2026-03-13 07:43:46.093723998 +0000 UTC m=+285.764502980" watchObservedRunningTime="2026-03-13 07:43:46.094922348 +0000 UTC m=+285.765701340" Mar 13 07:43:51 crc kubenswrapper[4876]: I0313 07:43:51.102469 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgzdt" event={"ID":"af6d5722-1354-4021-9ba4-6ad9a1286c60","Type":"ContainerStarted","Data":"e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a"} Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.118007 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q97b" event={"ID":"63f3a5db-0db1-4486-a982-39c60bda9dc4","Type":"ContainerStarted","Data":"df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92"} Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.120343 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcl5" event={"ID":"8c486d8f-87cf-42c0-9b8c-8f715755edc8","Type":"ContainerStarted","Data":"5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033"} Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.121914 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gkxcc" event={"ID":"68fd169c-1d47-4f5b-a528-201a1ea159a1","Type":"ContainerStarted","Data":"7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c"} Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.124533 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cf5vz" event={"ID":"ddbd46d5-05cd-4d71-b085-142b13687f63","Type":"ContainerStarted","Data":"1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570"} Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.127405 4876 generic.go:334] "Generic (PLEG): container finished" podID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerID="e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a" exitCode=0 Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.127459 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgzdt" event={"ID":"af6d5722-1354-4021-9ba4-6ad9a1286c60","Type":"ContainerDied","Data":"e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a"} Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.134768 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdxtz" event={"ID":"f7fbdc82-2033-4493-8f3a-ee17b88eb94e","Type":"ContainerStarted","Data":"385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463"} Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.934951 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f9775c99c-6vjh2"] Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.935183 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" podUID="d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" containerName="controller-manager" containerID="cri-o://7a6d79b5c29ad20927dfc833471e1600538e5c984105a902449bed11f065232b" gracePeriod=30 Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.968439 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6"] Mar 13 07:43:52 crc kubenswrapper[4876]: I0313 07:43:52.968904 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" podUID="60d8edcc-191c-4651-818f-54b6a324837f" containerName="route-controller-manager" containerID="cri-o://6232c2487036bf7a8f7aa71d5f759be41146432089422b0210201f88353f54a5" gracePeriod=30 Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.196713 4876 generic.go:334] "Generic (PLEG): container finished" podID="d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" containerID="7a6d79b5c29ad20927dfc833471e1600538e5c984105a902449bed11f065232b" exitCode=0 Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.196760 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" event={"ID":"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf","Type":"ContainerDied","Data":"7a6d79b5c29ad20927dfc833471e1600538e5c984105a902449bed11f065232b"} Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.198548 4876 generic.go:334] "Generic (PLEG): container finished" podID="60d8edcc-191c-4651-818f-54b6a324837f" containerID="6232c2487036bf7a8f7aa71d5f759be41146432089422b0210201f88353f54a5" exitCode=0 Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.198600 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" event={"ID":"60d8edcc-191c-4651-818f-54b6a324837f","Type":"ContainerDied","Data":"6232c2487036bf7a8f7aa71d5f759be41146432089422b0210201f88353f54a5"} Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.199999 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgzdt" event={"ID":"af6d5722-1354-4021-9ba4-6ad9a1286c60","Type":"ContainerStarted","Data":"6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7"} Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.202473 4876 generic.go:334] "Generic (PLEG): container finished" podID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerID="df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92" exitCode=0 Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.202496 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q97b" event={"ID":"63f3a5db-0db1-4486-a982-39c60bda9dc4","Type":"ContainerDied","Data":"df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92"} Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.225073 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kgzdt" podStartSLOduration=4.442095364 podStartE2EDuration="1m1.22505808s" podCreationTimestamp="2026-03-13 07:42:52 +0000 UTC" firstStartedPulling="2026-03-13 07:42:55.866994775 +0000 UTC m=+235.537773757" lastFinishedPulling="2026-03-13 07:43:52.649957491 +0000 UTC m=+292.320736473" observedRunningTime="2026-03-13 07:43:53.222047388 +0000 UTC m=+292.892826370" watchObservedRunningTime="2026-03-13 07:43:53.22505808 +0000 UTC m=+292.895837062" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.728198 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.728954 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.748327 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.787306 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.923217 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-proxy-ca-bundles\") pod \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.923572 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsrfz\" (UniqueName: \"kubernetes.io/projected/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-kube-api-access-wsrfz\") pod \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.923692 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fftk2\" (UniqueName: \"kubernetes.io/projected/60d8edcc-191c-4651-818f-54b6a324837f-kube-api-access-fftk2\") pod \"60d8edcc-191c-4651-818f-54b6a324837f\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.923780 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-serving-cert\") pod \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.923873 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-client-ca\") pod \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.924005 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-client-ca\") pod \"60d8edcc-191c-4651-818f-54b6a324837f\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.924071 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d8edcc-191c-4651-818f-54b6a324837f-serving-cert\") pod \"60d8edcc-191c-4651-818f-54b6a324837f\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.924148 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-config\") pod \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\" (UID: \"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.924214 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-config\") pod \"60d8edcc-191c-4651-818f-54b6a324837f\" (UID: \"60d8edcc-191c-4651-818f-54b6a324837f\") " Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.924556 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" (UID: "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.925052 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-client-ca" (OuterVolumeSpecName: "client-ca") pod "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" (UID: "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.927337 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-client-ca" (OuterVolumeSpecName: "client-ca") pod "60d8edcc-191c-4651-818f-54b6a324837f" (UID: "60d8edcc-191c-4651-818f-54b6a324837f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.927924 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-config" (OuterVolumeSpecName: "config") pod "60d8edcc-191c-4651-818f-54b6a324837f" (UID: "60d8edcc-191c-4651-818f-54b6a324837f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.927943 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-config" (OuterVolumeSpecName: "config") pod "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" (UID: "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.947759 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-kube-api-access-wsrfz" (OuterVolumeSpecName: "kube-api-access-wsrfz") pod "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" (UID: "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf"). InnerVolumeSpecName "kube-api-access-wsrfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.953834 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60d8edcc-191c-4651-818f-54b6a324837f-kube-api-access-fftk2" (OuterVolumeSpecName: "kube-api-access-fftk2") pod "60d8edcc-191c-4651-818f-54b6a324837f" (UID: "60d8edcc-191c-4651-818f-54b6a324837f"). InnerVolumeSpecName "kube-api-access-fftk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.955435 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" (UID: "d32a8720-cb92-4f9b-8cbb-7feeffc07ebf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:43:53 crc kubenswrapper[4876]: I0313 07:43:53.964014 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d8edcc-191c-4651-818f-54b6a324837f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "60d8edcc-191c-4651-818f-54b6a324837f" (UID: "60d8edcc-191c-4651-818f-54b6a324837f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026052 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026090 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d8edcc-191c-4651-818f-54b6a324837f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026102 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026114 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d8edcc-191c-4651-818f-54b6a324837f-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026125 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026140 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsrfz\" (UniqueName: \"kubernetes.io/projected/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-kube-api-access-wsrfz\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026153 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fftk2\" (UniqueName: \"kubernetes.io/projected/60d8edcc-191c-4651-818f-54b6a324837f-kube-api-access-fftk2\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026165 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.026178 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.214486 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" event={"ID":"60d8edcc-191c-4651-818f-54b6a324837f","Type":"ContainerDied","Data":"1d8093014768ff7ae73a183e0f41c47551bad9d9b665f4e9eb5ccad826f5cecd"} Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.214539 4876 scope.go:117] "RemoveContainer" containerID="6232c2487036bf7a8f7aa71d5f759be41146432089422b0210201f88353f54a5" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.214688 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.222805 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bg8rq" event={"ID":"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1","Type":"ContainerStarted","Data":"441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe"} Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.235637 4876 generic.go:334] "Generic (PLEG): container finished" podID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerID="1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570" exitCode=0 Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.235691 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cf5vz" event={"ID":"ddbd46d5-05cd-4d71-b085-142b13687f63","Type":"ContainerDied","Data":"1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570"} Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.239376 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q97b" event={"ID":"63f3a5db-0db1-4486-a982-39c60bda9dc4","Type":"ContainerStarted","Data":"24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346"} Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.241532 4876 generic.go:334] "Generic (PLEG): container finished" podID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerID="5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033" exitCode=0 Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.241659 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcl5" event={"ID":"8c486d8f-87cf-42c0-9b8c-8f715755edc8","Type":"ContainerDied","Data":"5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033"} Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.244024 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" event={"ID":"d32a8720-cb92-4f9b-8cbb-7feeffc07ebf","Type":"ContainerDied","Data":"bbc51cc8c7a2c8a73973bd5e5c3b53719735da18fada393e13434779a4bf1918"} Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.244068 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f9775c99c-6vjh2" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.248071 4876 scope.go:117] "RemoveContainer" containerID="7a6d79b5c29ad20927dfc833471e1600538e5c984105a902449bed11f065232b" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.287674 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6"] Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.313506 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-756688789b-z9ml6"] Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.321692 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.323488 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.348031 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2q97b" podStartSLOduration=3.613266802 podStartE2EDuration="1m0.348016277s" podCreationTimestamp="2026-03-13 07:42:54 +0000 UTC" firstStartedPulling="2026-03-13 07:42:56.954195571 +0000 UTC m=+236.624974553" lastFinishedPulling="2026-03-13 07:43:53.688945046 +0000 UTC m=+293.359724028" observedRunningTime="2026-03-13 07:43:54.344876422 +0000 UTC m=+294.015655404" watchObservedRunningTime="2026-03-13 07:43:54.348016277 +0000 UTC m=+294.018795259" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.352161 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.405607 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f9775c99c-6vjh2"] Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.412587 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5f9775c99c-6vjh2"] Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.607389 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.607435 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.607477 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.607970 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.608015 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c" gracePeriod=600 Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.964373 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5c988b6444-tdqj6"] Mar 13 07:43:54 crc kubenswrapper[4876]: E0313 07:43:54.964726 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d8edcc-191c-4651-818f-54b6a324837f" containerName="route-controller-manager" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.964746 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d8edcc-191c-4651-818f-54b6a324837f" containerName="route-controller-manager" Mar 13 07:43:54 crc kubenswrapper[4876]: E0313 07:43:54.964765 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" containerName="controller-manager" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.964772 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" containerName="controller-manager" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.964874 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" containerName="controller-manager" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.964885 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="60d8edcc-191c-4651-818f-54b6a324837f" containerName="route-controller-manager" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.965392 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.968768 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c"] Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.969694 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.971849 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.971876 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.972038 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.972111 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.972376 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.972710 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.972863 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.972990 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.973139 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.973263 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.973459 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.976763 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.980818 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 07:43:54 crc kubenswrapper[4876]: I0313 07:43:54.992797 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c"] Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046416 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60d8edcc-191c-4651-818f-54b6a324837f" path="/var/lib/kubelet/pods/60d8edcc-191c-4651-818f-54b6a324837f/volumes" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046526 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-client-ca\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046561 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r75sb\" (UniqueName: \"kubernetes.io/projected/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-kube-api-access-r75sb\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046590 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-client-ca\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046606 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a66d689-db59-4913-8e1f-d108bf1fe370-serving-cert\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046620 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-config\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046663 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzr2s\" (UniqueName: \"kubernetes.io/projected/4a66d689-db59-4913-8e1f-d108bf1fe370-kube-api-access-rzr2s\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046681 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-serving-cert\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046715 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-proxy-ca-bundles\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046737 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-config\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.046991 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d32a8720-cb92-4f9b-8cbb-7feeffc07ebf" path="/var/lib/kubelet/pods/d32a8720-cb92-4f9b-8cbb-7feeffc07ebf/volumes" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.054900 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c988b6444-tdqj6"] Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149195 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-client-ca\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149529 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-config\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149549 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a66d689-db59-4913-8e1f-d108bf1fe370-serving-cert\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzr2s\" (UniqueName: \"kubernetes.io/projected/4a66d689-db59-4913-8e1f-d108bf1fe370-kube-api-access-rzr2s\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149607 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-serving-cert\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149641 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-proxy-ca-bundles\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149663 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-config\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149706 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-client-ca\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.149725 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r75sb\" (UniqueName: \"kubernetes.io/projected/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-kube-api-access-r75sb\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.151766 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-client-ca\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.152104 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-config\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.153157 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-client-ca\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.153516 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-proxy-ca-bundles\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.154010 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-config\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.165019 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a66d689-db59-4913-8e1f-d108bf1fe370-serving-cert\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.165038 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-serving-cert\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.177396 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzr2s\" (UniqueName: \"kubernetes.io/projected/4a66d689-db59-4913-8e1f-d108bf1fe370-kube-api-access-rzr2s\") pod \"controller-manager-5c988b6444-tdqj6\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.183327 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r75sb\" (UniqueName: \"kubernetes.io/projected/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-kube-api-access-r75sb\") pod \"route-controller-manager-6999bfff6-6kw2c\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.253391 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c" exitCode=0 Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.253468 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c"} Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.254819 4876 generic.go:334] "Generic (PLEG): container finished" podID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerID="385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463" exitCode=0 Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.254861 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdxtz" event={"ID":"f7fbdc82-2033-4493-8f3a-ee17b88eb94e","Type":"ContainerDied","Data":"385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463"} Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.294976 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.297233 4876 generic.go:334] "Generic (PLEG): container finished" podID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerID="441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe" exitCode=0 Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.297427 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bg8rq" event={"ID":"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1","Type":"ContainerDied","Data":"441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe"} Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.301895 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.314185 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6cnvw" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.377681 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.377726 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.408655 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kgzdt" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="registry-server" probeResult="failure" output=< Mar 13 07:43:55 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 07:43:55 crc kubenswrapper[4876]: > Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.426629 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.672380 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c"] Mar 13 07:43:55 crc kubenswrapper[4876]: I0313 07:43:55.729457 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c988b6444-tdqj6"] Mar 13 07:43:56 crc kubenswrapper[4876]: I0313 07:43:56.303280 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" event={"ID":"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f","Type":"ContainerStarted","Data":"0aeb5f7210246c2de9fc8e4e18f5480a980406b46f16dea10a0b81079fdfebc8"} Mar 13 07:43:56 crc kubenswrapper[4876]: I0313 07:43:56.304067 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" event={"ID":"4a66d689-db59-4913-8e1f-d108bf1fe370","Type":"ContainerStarted","Data":"9fe12112194182acff6abd45c73b85933bf552778d6b16dae80ec9bfda38d3e9"} Mar 13 07:43:56 crc kubenswrapper[4876]: I0313 07:43:56.306025 4876 generic.go:334] "Generic (PLEG): container finished" podID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerID="7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c" exitCode=0 Mar 13 07:43:56 crc kubenswrapper[4876]: I0313 07:43:56.306085 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gkxcc" event={"ID":"68fd169c-1d47-4f5b-a528-201a1ea159a1","Type":"ContainerDied","Data":"7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c"} Mar 13 07:43:56 crc kubenswrapper[4876]: I0313 07:43:56.471895 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-2q97b" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="registry-server" probeResult="failure" output=< Mar 13 07:43:56 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 07:43:56 crc kubenswrapper[4876]: > Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.337689 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdxtz" event={"ID":"f7fbdc82-2033-4493-8f3a-ee17b88eb94e","Type":"ContainerStarted","Data":"157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473"} Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.342801 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcl5" event={"ID":"8c486d8f-87cf-42c0-9b8c-8f715755edc8","Type":"ContainerStarted","Data":"feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38"} Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.344952 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" event={"ID":"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f","Type":"ContainerStarted","Data":"dd88861adab01940a851ebcd8cbefba44db0e5bbadef910dded117593827a74f"} Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.345170 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.346943 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" event={"ID":"4a66d689-db59-4913-8e1f-d108bf1fe370","Type":"ContainerStarted","Data":"d3da9ca0e252d4e162916f4ec0c853dbe9a056e1f286a15372d3a05b319a9ddb"} Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.347164 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.349779 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bg8rq" event={"ID":"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1","Type":"ContainerStarted","Data":"e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0"} Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.352021 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cf5vz" event={"ID":"ddbd46d5-05cd-4d71-b085-142b13687f63","Type":"ContainerStarted","Data":"df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89"} Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.354188 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"9f60a1005aa99b7a7a6f69d8ddd25eddf167d10b844bb62bc7097f7905502e52"} Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.355019 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.357458 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.361877 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kdxtz" podStartSLOduration=3.688214058 podStartE2EDuration="1m2.361858006s" podCreationTimestamp="2026-03-13 07:42:55 +0000 UTC" firstStartedPulling="2026-03-13 07:42:58.136641683 +0000 UTC m=+237.807420665" lastFinishedPulling="2026-03-13 07:43:56.810285631 +0000 UTC m=+296.481064613" observedRunningTime="2026-03-13 07:43:57.360978356 +0000 UTC m=+297.031757358" watchObservedRunningTime="2026-03-13 07:43:57.361858006 +0000 UTC m=+297.032636998" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.379692 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bg8rq" podStartSLOduration=3.5852903879999998 podStartE2EDuration="1m2.379675536s" podCreationTimestamp="2026-03-13 07:42:55 +0000 UTC" firstStartedPulling="2026-03-13 07:42:58.125741627 +0000 UTC m=+237.796520609" lastFinishedPulling="2026-03-13 07:43:56.920126785 +0000 UTC m=+296.590905757" observedRunningTime="2026-03-13 07:43:57.377026656 +0000 UTC m=+297.047805638" watchObservedRunningTime="2026-03-13 07:43:57.379675536 +0000 UTC m=+297.050454518" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.397115 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" podStartSLOduration=5.397098764 podStartE2EDuration="5.397098764s" podCreationTimestamp="2026-03-13 07:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:43:57.395231241 +0000 UTC m=+297.066010223" watchObservedRunningTime="2026-03-13 07:43:57.397098764 +0000 UTC m=+297.067877746" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.459666 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cf5vz" podStartSLOduration=3.6319646089999997 podStartE2EDuration="1m4.459644583s" podCreationTimestamp="2026-03-13 07:42:53 +0000 UTC" firstStartedPulling="2026-03-13 07:42:55.866544592 +0000 UTC m=+235.537323574" lastFinishedPulling="2026-03-13 07:43:56.694224566 +0000 UTC m=+296.365003548" observedRunningTime="2026-03-13 07:43:57.439657689 +0000 UTC m=+297.110436671" watchObservedRunningTime="2026-03-13 07:43:57.459644583 +0000 UTC m=+297.130423595" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.479353 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fmcl5" podStartSLOduration=4.640002597 podStartE2EDuration="1m5.479338187s" podCreationTimestamp="2026-03-13 07:42:52 +0000 UTC" firstStartedPulling="2026-03-13 07:42:55.867412348 +0000 UTC m=+235.538191330" lastFinishedPulling="2026-03-13 07:43:56.706747928 +0000 UTC m=+296.377526920" observedRunningTime="2026-03-13 07:43:57.476663838 +0000 UTC m=+297.147442840" watchObservedRunningTime="2026-03-13 07:43:57.479338187 +0000 UTC m=+297.150117169" Mar 13 07:43:57 crc kubenswrapper[4876]: I0313 07:43:57.479495 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" podStartSLOduration=5.479491573 podStartE2EDuration="5.479491573s" podCreationTimestamp="2026-03-13 07:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:43:57.458346399 +0000 UTC m=+297.129125391" watchObservedRunningTime="2026-03-13 07:43:57.479491573 +0000 UTC m=+297.150270545" Mar 13 07:43:58 crc kubenswrapper[4876]: I0313 07:43:58.361013 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gkxcc" event={"ID":"68fd169c-1d47-4f5b-a528-201a1ea159a1","Type":"ContainerStarted","Data":"702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192"} Mar 13 07:43:58 crc kubenswrapper[4876]: I0313 07:43:58.386606 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gkxcc" podStartSLOduration=2.607143824 podStartE2EDuration="1m2.386584729s" podCreationTimestamp="2026-03-13 07:42:56 +0000 UTC" firstStartedPulling="2026-03-13 07:42:58.076552607 +0000 UTC m=+237.747331589" lastFinishedPulling="2026-03-13 07:43:57.855993522 +0000 UTC m=+297.526772494" observedRunningTime="2026-03-13 07:43:58.385029937 +0000 UTC m=+298.055808919" watchObservedRunningTime="2026-03-13 07:43:58.386584729 +0000 UTC m=+298.057363711" Mar 13 07:43:59 crc kubenswrapper[4876]: I0313 07:43:59.471017 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mtcn4"] Mar 13 07:43:59 crc kubenswrapper[4876]: I0313 07:43:59.471850 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mtcn4" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerName="registry-server" containerID="cri-o://9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b" gracePeriod=2 Mar 13 07:43:59 crc kubenswrapper[4876]: I0313 07:43:59.888526 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.022050 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrvnp\" (UniqueName: \"kubernetes.io/projected/18c1ffa5-3580-48f1-8df2-5f442d47062d-kube-api-access-zrvnp\") pod \"18c1ffa5-3580-48f1-8df2-5f442d47062d\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.022158 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-catalog-content\") pod \"18c1ffa5-3580-48f1-8df2-5f442d47062d\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.022203 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-utilities\") pod \"18c1ffa5-3580-48f1-8df2-5f442d47062d\" (UID: \"18c1ffa5-3580-48f1-8df2-5f442d47062d\") " Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.023088 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-utilities" (OuterVolumeSpecName: "utilities") pod "18c1ffa5-3580-48f1-8df2-5f442d47062d" (UID: "18c1ffa5-3580-48f1-8df2-5f442d47062d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.030461 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18c1ffa5-3580-48f1-8df2-5f442d47062d-kube-api-access-zrvnp" (OuterVolumeSpecName: "kube-api-access-zrvnp") pod "18c1ffa5-3580-48f1-8df2-5f442d47062d" (UID: "18c1ffa5-3580-48f1-8df2-5f442d47062d"). InnerVolumeSpecName "kube-api-access-zrvnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.077970 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18c1ffa5-3580-48f1-8df2-5f442d47062d" (UID: "18c1ffa5-3580-48f1-8df2-5f442d47062d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.123879 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.123925 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrvnp\" (UniqueName: \"kubernetes.io/projected/18c1ffa5-3580-48f1-8df2-5f442d47062d-kube-api-access-zrvnp\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.123941 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18c1ffa5-3580-48f1-8df2-5f442d47062d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.136746 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556464-ctthh"] Mar 13 07:44:00 crc kubenswrapper[4876]: E0313 07:44:00.137002 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerName="registry-server" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.137019 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerName="registry-server" Mar 13 07:44:00 crc kubenswrapper[4876]: E0313 07:44:00.137032 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerName="extract-content" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.137040 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerName="extract-content" Mar 13 07:44:00 crc kubenswrapper[4876]: E0313 07:44:00.137063 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerName="extract-utilities" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.137072 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerName="extract-utilities" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.137200 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerName="registry-server" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.137687 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556464-ctthh" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.139667 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.139906 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.141408 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.149525 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556464-ctthh"] Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.326130 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjqf8\" (UniqueName: \"kubernetes.io/projected/134014ac-2aa1-411c-a5af-492088dee314-kube-api-access-zjqf8\") pod \"auto-csr-approver-29556464-ctthh\" (UID: \"134014ac-2aa1-411c-a5af-492088dee314\") " pod="openshift-infra/auto-csr-approver-29556464-ctthh" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.375996 4876 generic.go:334] "Generic (PLEG): container finished" podID="18c1ffa5-3580-48f1-8df2-5f442d47062d" containerID="9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b" exitCode=0 Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.376053 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtcn4" event={"ID":"18c1ffa5-3580-48f1-8df2-5f442d47062d","Type":"ContainerDied","Data":"9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b"} Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.376061 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtcn4" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.376092 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtcn4" event={"ID":"18c1ffa5-3580-48f1-8df2-5f442d47062d","Type":"ContainerDied","Data":"fcd55685e2d927ed8f2ddbb66a591fe88507a2d538d334fa47c422fcf5092a2f"} Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.376117 4876 scope.go:117] "RemoveContainer" containerID="9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.405530 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mtcn4"] Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.406727 4876 scope.go:117] "RemoveContainer" containerID="8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.408290 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mtcn4"] Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.422552 4876 scope.go:117] "RemoveContainer" containerID="4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.428262 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjqf8\" (UniqueName: \"kubernetes.io/projected/134014ac-2aa1-411c-a5af-492088dee314-kube-api-access-zjqf8\") pod \"auto-csr-approver-29556464-ctthh\" (UID: \"134014ac-2aa1-411c-a5af-492088dee314\") " pod="openshift-infra/auto-csr-approver-29556464-ctthh" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.442616 4876 scope.go:117] "RemoveContainer" containerID="9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b" Mar 13 07:44:00 crc kubenswrapper[4876]: E0313 07:44:00.443128 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b\": container with ID starting with 9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b not found: ID does not exist" containerID="9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.443181 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b"} err="failed to get container status \"9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b\": rpc error: code = NotFound desc = could not find container \"9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b\": container with ID starting with 9d18b2929184898b3f9da15e1431ed9e35e6b2d8d003d9142e38e4767abd783b not found: ID does not exist" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.443211 4876 scope.go:117] "RemoveContainer" containerID="8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7" Mar 13 07:44:00 crc kubenswrapper[4876]: E0313 07:44:00.443740 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7\": container with ID starting with 8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7 not found: ID does not exist" containerID="8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.443821 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7"} err="failed to get container status \"8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7\": rpc error: code = NotFound desc = could not find container \"8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7\": container with ID starting with 8e28bef734a67e7d46224cdb47f69dffac50f02a7e21561a98f3689fa50ee2c7 not found: ID does not exist" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.443850 4876 scope.go:117] "RemoveContainer" containerID="4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94" Mar 13 07:44:00 crc kubenswrapper[4876]: E0313 07:44:00.444834 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94\": container with ID starting with 4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94 not found: ID does not exist" containerID="4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.444861 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94"} err="failed to get container status \"4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94\": rpc error: code = NotFound desc = could not find container \"4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94\": container with ID starting with 4e9b1d21060b6390ee047c38a3e7267dd22d5e2963507f9b500f1198f6effe94 not found: ID does not exist" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.448130 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjqf8\" (UniqueName: \"kubernetes.io/projected/134014ac-2aa1-411c-a5af-492088dee314-kube-api-access-zjqf8\") pod \"auto-csr-approver-29556464-ctthh\" (UID: \"134014ac-2aa1-411c-a5af-492088dee314\") " pod="openshift-infra/auto-csr-approver-29556464-ctthh" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.454388 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556464-ctthh" Mar 13 07:44:00 crc kubenswrapper[4876]: I0313 07:44:00.891148 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556464-ctthh"] Mar 13 07:44:01 crc kubenswrapper[4876]: I0313 07:44:01.043021 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18c1ffa5-3580-48f1-8df2-5f442d47062d" path="/var/lib/kubelet/pods/18c1ffa5-3580-48f1-8df2-5f442d47062d/volumes" Mar 13 07:44:01 crc kubenswrapper[4876]: I0313 07:44:01.382056 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556464-ctthh" event={"ID":"134014ac-2aa1-411c-a5af-492088dee314","Type":"ContainerStarted","Data":"f789cc1f7da3d8be2cfe0b736ec8b12cd1beebf9c14d77a83df1a3005305cf2e"} Mar 13 07:44:03 crc kubenswrapper[4876]: I0313 07:44:03.290861 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:44:03 crc kubenswrapper[4876]: I0313 07:44:03.291864 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:44:03 crc kubenswrapper[4876]: I0313 07:44:03.343541 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:44:03 crc kubenswrapper[4876]: I0313 07:44:03.437101 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:44:04 crc kubenswrapper[4876]: I0313 07:44:04.321520 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:44:04 crc kubenswrapper[4876]: I0313 07:44:04.321853 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:44:04 crc kubenswrapper[4876]: I0313 07:44:04.364467 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:44:04 crc kubenswrapper[4876]: I0313 07:44:04.376969 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:44:04 crc kubenswrapper[4876]: I0313 07:44:04.412649 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:44:04 crc kubenswrapper[4876]: I0313 07:44:04.445960 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:44:05 crc kubenswrapper[4876]: I0313 07:44:05.406612 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556464-ctthh" event={"ID":"134014ac-2aa1-411c-a5af-492088dee314","Type":"ContainerStarted","Data":"e98d759f76ecebdc8f337c52e8649cb616f630db684e7ba0690a4b677228d42e"} Mar 13 07:44:05 crc kubenswrapper[4876]: I0313 07:44:05.422314 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556464-ctthh" podStartSLOduration=1.182525927 podStartE2EDuration="5.422293446s" podCreationTimestamp="2026-03-13 07:44:00 +0000 UTC" firstStartedPulling="2026-03-13 07:44:00.912088905 +0000 UTC m=+300.582867897" lastFinishedPulling="2026-03-13 07:44:05.151856434 +0000 UTC m=+304.822635416" observedRunningTime="2026-03-13 07:44:05.418913652 +0000 UTC m=+305.089692634" watchObservedRunningTime="2026-03-13 07:44:05.422293446 +0000 UTC m=+305.093072438" Mar 13 07:44:05 crc kubenswrapper[4876]: I0313 07:44:05.426645 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:44:05 crc kubenswrapper[4876]: I0313 07:44:05.470375 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:44:05 crc kubenswrapper[4876]: I0313 07:44:05.805951 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:44:05 crc kubenswrapper[4876]: I0313 07:44:05.806053 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:44:05 crc kubenswrapper[4876]: I0313 07:44:05.853483 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.417470 4876 generic.go:334] "Generic (PLEG): container finished" podID="134014ac-2aa1-411c-a5af-492088dee314" containerID="e98d759f76ecebdc8f337c52e8649cb616f630db684e7ba0690a4b677228d42e" exitCode=0 Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.417577 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556464-ctthh" event={"ID":"134014ac-2aa1-411c-a5af-492088dee314","Type":"ContainerDied","Data":"e98d759f76ecebdc8f337c52e8649cb616f630db684e7ba0690a4b677228d42e"} Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.474533 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.474764 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.476962 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.521717 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.650032 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.650113 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.673928 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cf5vz"] Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.674334 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cf5vz" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerName="registry-server" containerID="cri-o://df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89" gracePeriod=2 Mar 13 07:44:06 crc kubenswrapper[4876]: I0313 07:44:06.706695 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.177345 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.325166 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-utilities\") pod \"ddbd46d5-05cd-4d71-b085-142b13687f63\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.325336 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-catalog-content\") pod \"ddbd46d5-05cd-4d71-b085-142b13687f63\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.325391 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr79z\" (UniqueName: \"kubernetes.io/projected/ddbd46d5-05cd-4d71-b085-142b13687f63-kube-api-access-rr79z\") pod \"ddbd46d5-05cd-4d71-b085-142b13687f63\" (UID: \"ddbd46d5-05cd-4d71-b085-142b13687f63\") " Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.327204 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-utilities" (OuterVolumeSpecName: "utilities") pod "ddbd46d5-05cd-4d71-b085-142b13687f63" (UID: "ddbd46d5-05cd-4d71-b085-142b13687f63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.333530 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddbd46d5-05cd-4d71-b085-142b13687f63-kube-api-access-rr79z" (OuterVolumeSpecName: "kube-api-access-rr79z") pod "ddbd46d5-05cd-4d71-b085-142b13687f63" (UID: "ddbd46d5-05cd-4d71-b085-142b13687f63"). InnerVolumeSpecName "kube-api-access-rr79z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.385568 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ddbd46d5-05cd-4d71-b085-142b13687f63" (UID: "ddbd46d5-05cd-4d71-b085-142b13687f63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.427739 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.427796 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddbd46d5-05cd-4d71-b085-142b13687f63-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.427816 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr79z\" (UniqueName: \"kubernetes.io/projected/ddbd46d5-05cd-4d71-b085-142b13687f63-kube-api-access-rr79z\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.430035 4876 generic.go:334] "Generic (PLEG): container finished" podID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerID="df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89" exitCode=0 Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.430090 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cf5vz" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.430135 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cf5vz" event={"ID":"ddbd46d5-05cd-4d71-b085-142b13687f63","Type":"ContainerDied","Data":"df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89"} Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.430195 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cf5vz" event={"ID":"ddbd46d5-05cd-4d71-b085-142b13687f63","Type":"ContainerDied","Data":"59d822d7a5fdea9cd1e2e79fa2183c85fea8e238f88747f59d4dc07b60226db5"} Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.430218 4876 scope.go:117] "RemoveContainer" containerID="df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.452422 4876 scope.go:117] "RemoveContainer" containerID="1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.471183 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cf5vz"] Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.474631 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cf5vz"] Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.487910 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.489802 4876 scope.go:117] "RemoveContainer" containerID="3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.490543 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.510639 4876 scope.go:117] "RemoveContainer" containerID="df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89" Mar 13 07:44:07 crc kubenswrapper[4876]: E0313 07:44:07.512094 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89\": container with ID starting with df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89 not found: ID does not exist" containerID="df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.512162 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89"} err="failed to get container status \"df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89\": rpc error: code = NotFound desc = could not find container \"df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89\": container with ID starting with df64547da8c29faed60d9d4ed136009da31e049baa0f154c9728a219f18fda89 not found: ID does not exist" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.512207 4876 scope.go:117] "RemoveContainer" containerID="1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570" Mar 13 07:44:07 crc kubenswrapper[4876]: E0313 07:44:07.512871 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570\": container with ID starting with 1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570 not found: ID does not exist" containerID="1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.512906 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570"} err="failed to get container status \"1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570\": rpc error: code = NotFound desc = could not find container \"1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570\": container with ID starting with 1c6471c9acde737f3e348333e2e22fcb35db6b8726dc1998ffb0284681197570 not found: ID does not exist" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.512931 4876 scope.go:117] "RemoveContainer" containerID="3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33" Mar 13 07:44:07 crc kubenswrapper[4876]: E0313 07:44:07.513291 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33\": container with ID starting with 3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33 not found: ID does not exist" containerID="3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.513313 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33"} err="failed to get container status \"3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33\": rpc error: code = NotFound desc = could not find container \"3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33\": container with ID starting with 3926712696eb4b577126d8a098bbad322b24462ac9ae4930b338f2574d30de33 not found: ID does not exist" Mar 13 07:44:07 crc kubenswrapper[4876]: I0313 07:44:07.850933 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556464-ctthh" Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.042591 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjqf8\" (UniqueName: \"kubernetes.io/projected/134014ac-2aa1-411c-a5af-492088dee314-kube-api-access-zjqf8\") pod \"134014ac-2aa1-411c-a5af-492088dee314\" (UID: \"134014ac-2aa1-411c-a5af-492088dee314\") " Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.048424 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/134014ac-2aa1-411c-a5af-492088dee314-kube-api-access-zjqf8" (OuterVolumeSpecName: "kube-api-access-zjqf8") pod "134014ac-2aa1-411c-a5af-492088dee314" (UID: "134014ac-2aa1-411c-a5af-492088dee314"). InnerVolumeSpecName "kube-api-access-zjqf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.145104 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjqf8\" (UniqueName: \"kubernetes.io/projected/134014ac-2aa1-411c-a5af-492088dee314-kube-api-access-zjqf8\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.438893 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556464-ctthh" event={"ID":"134014ac-2aa1-411c-a5af-492088dee314","Type":"ContainerDied","Data":"f789cc1f7da3d8be2cfe0b736ec8b12cd1beebf9c14d77a83df1a3005305cf2e"} Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.439584 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f789cc1f7da3d8be2cfe0b736ec8b12cd1beebf9c14d77a83df1a3005305cf2e" Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.438997 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556464-ctthh" Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.470607 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bg8rq"] Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.471143 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bg8rq" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerName="registry-server" containerID="cri-o://e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0" gracePeriod=2 Mar 13 07:44:08 crc kubenswrapper[4876]: I0313 07:44:08.929824 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.047052 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" path="/var/lib/kubelet/pods/ddbd46d5-05cd-4d71-b085-142b13687f63/volumes" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.060625 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-catalog-content\") pod \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.061388 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-utilities\") pod \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.061619 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5lkh\" (UniqueName: \"kubernetes.io/projected/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-kube-api-access-r5lkh\") pod \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\" (UID: \"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1\") " Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.062555 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-utilities" (OuterVolumeSpecName: "utilities") pod "c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" (UID: "c4259aab-8c71-4b40-957e-ef4e5ddc1ea1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.074501 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-kube-api-access-r5lkh" (OuterVolumeSpecName: "kube-api-access-r5lkh") pod "c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" (UID: "c4259aab-8c71-4b40-957e-ef4e5ddc1ea1"). InnerVolumeSpecName "kube-api-access-r5lkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.083637 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gkxcc"] Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.106355 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" (UID: "c4259aab-8c71-4b40-957e-ef4e5ddc1ea1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.163223 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.163328 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5lkh\" (UniqueName: \"kubernetes.io/projected/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-kube-api-access-r5lkh\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.163352 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.452510 4876 generic.go:334] "Generic (PLEG): container finished" podID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerID="e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0" exitCode=0 Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.452668 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bg8rq" event={"ID":"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1","Type":"ContainerDied","Data":"e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0"} Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.452717 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bg8rq" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.452773 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bg8rq" event={"ID":"c4259aab-8c71-4b40-957e-ef4e5ddc1ea1","Type":"ContainerDied","Data":"faaa89fa9f550f216225d2c0149e2ff2ec70f09230a7f549797faddfb647c1b4"} Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.452797 4876 scope.go:117] "RemoveContainer" containerID="e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.453096 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gkxcc" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerName="registry-server" containerID="cri-o://702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192" gracePeriod=2 Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.490350 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bg8rq"] Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.493621 4876 scope.go:117] "RemoveContainer" containerID="441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.494113 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bg8rq"] Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.513322 4876 scope.go:117] "RemoveContainer" containerID="72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.603985 4876 scope.go:117] "RemoveContainer" containerID="e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0" Mar 13 07:44:09 crc kubenswrapper[4876]: E0313 07:44:09.604651 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0\": container with ID starting with e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0 not found: ID does not exist" containerID="e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.604698 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0"} err="failed to get container status \"e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0\": rpc error: code = NotFound desc = could not find container \"e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0\": container with ID starting with e507e9ea5e37783810de6be6b3a4983389d10b5103fad71788404000117cc9d0 not found: ID does not exist" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.604742 4876 scope.go:117] "RemoveContainer" containerID="441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe" Mar 13 07:44:09 crc kubenswrapper[4876]: E0313 07:44:09.605195 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe\": container with ID starting with 441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe not found: ID does not exist" containerID="441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.605224 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe"} err="failed to get container status \"441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe\": rpc error: code = NotFound desc = could not find container \"441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe\": container with ID starting with 441ddb8416a45a676147d7563881d44fbb11302aafe0f5dbd5bcd9132fd168fe not found: ID does not exist" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.605259 4876 scope.go:117] "RemoveContainer" containerID="72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44" Mar 13 07:44:09 crc kubenswrapper[4876]: E0313 07:44:09.605669 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44\": container with ID starting with 72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44 not found: ID does not exist" containerID="72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.605696 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44"} err="failed to get container status \"72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44\": rpc error: code = NotFound desc = could not find container \"72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44\": container with ID starting with 72ef7b9314ef7732bb767079a52db3f99973a35fa54d8aa8a54e11fb09beff44 not found: ID does not exist" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.906103 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.975488 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7bxl\" (UniqueName: \"kubernetes.io/projected/68fd169c-1d47-4f5b-a528-201a1ea159a1-kube-api-access-v7bxl\") pod \"68fd169c-1d47-4f5b-a528-201a1ea159a1\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.975561 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-utilities\") pod \"68fd169c-1d47-4f5b-a528-201a1ea159a1\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.975607 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-catalog-content\") pod \"68fd169c-1d47-4f5b-a528-201a1ea159a1\" (UID: \"68fd169c-1d47-4f5b-a528-201a1ea159a1\") " Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.977569 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-utilities" (OuterVolumeSpecName: "utilities") pod "68fd169c-1d47-4f5b-a528-201a1ea159a1" (UID: "68fd169c-1d47-4f5b-a528-201a1ea159a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:44:09 crc kubenswrapper[4876]: I0313 07:44:09.984766 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68fd169c-1d47-4f5b-a528-201a1ea159a1-kube-api-access-v7bxl" (OuterVolumeSpecName: "kube-api-access-v7bxl") pod "68fd169c-1d47-4f5b-a528-201a1ea159a1" (UID: "68fd169c-1d47-4f5b-a528-201a1ea159a1"). InnerVolumeSpecName "kube-api-access-v7bxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.078467 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7bxl\" (UniqueName: \"kubernetes.io/projected/68fd169c-1d47-4f5b-a528-201a1ea159a1-kube-api-access-v7bxl\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.078542 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.132744 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68fd169c-1d47-4f5b-a528-201a1ea159a1" (UID: "68fd169c-1d47-4f5b-a528-201a1ea159a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.180046 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fd169c-1d47-4f5b-a528-201a1ea159a1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.469480 4876 generic.go:334] "Generic (PLEG): container finished" podID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerID="702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192" exitCode=0 Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.469590 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gkxcc" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.469577 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gkxcc" event={"ID":"68fd169c-1d47-4f5b-a528-201a1ea159a1","Type":"ContainerDied","Data":"702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192"} Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.470554 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gkxcc" event={"ID":"68fd169c-1d47-4f5b-a528-201a1ea159a1","Type":"ContainerDied","Data":"a8a2f6aadef356e01fcb36bd4d8bd6294768da7386fbf3a366f125291e44d7f0"} Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.470590 4876 scope.go:117] "RemoveContainer" containerID="702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.499231 4876 scope.go:117] "RemoveContainer" containerID="7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.513574 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gkxcc"] Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.518477 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gkxcc"] Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.531352 4876 scope.go:117] "RemoveContainer" containerID="bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.553949 4876 scope.go:117] "RemoveContainer" containerID="702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192" Mar 13 07:44:10 crc kubenswrapper[4876]: E0313 07:44:10.554969 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192\": container with ID starting with 702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192 not found: ID does not exist" containerID="702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.555062 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192"} err="failed to get container status \"702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192\": rpc error: code = NotFound desc = could not find container \"702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192\": container with ID starting with 702c4dfcff341fa7032052916fc9de9fd7a45c0e090a93ed32b680017606e192 not found: ID does not exist" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.555112 4876 scope.go:117] "RemoveContainer" containerID="7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c" Mar 13 07:44:10 crc kubenswrapper[4876]: E0313 07:44:10.555855 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c\": container with ID starting with 7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c not found: ID does not exist" containerID="7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.555896 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c"} err="failed to get container status \"7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c\": rpc error: code = NotFound desc = could not find container \"7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c\": container with ID starting with 7747cca9875aab7f7602e13bf8ef6bb2059474b8214e08ac0bd626084336812c not found: ID does not exist" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.555930 4876 scope.go:117] "RemoveContainer" containerID="bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137" Mar 13 07:44:10 crc kubenswrapper[4876]: E0313 07:44:10.556453 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137\": container with ID starting with bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137 not found: ID does not exist" containerID="bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137" Mar 13 07:44:10 crc kubenswrapper[4876]: I0313 07:44:10.556509 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137"} err="failed to get container status \"bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137\": rpc error: code = NotFound desc = could not find container \"bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137\": container with ID starting with bf2460584e58605efd9551472afa2e782f2ea07e8c7c1c1ae960bce57448e137 not found: ID does not exist" Mar 13 07:44:11 crc kubenswrapper[4876]: I0313 07:44:11.043318 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" path="/var/lib/kubelet/pods/68fd169c-1d47-4f5b-a528-201a1ea159a1/volumes" Mar 13 07:44:11 crc kubenswrapper[4876]: I0313 07:44:11.044198 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" path="/var/lib/kubelet/pods/c4259aab-8c71-4b40-957e-ef4e5ddc1ea1/volumes" Mar 13 07:44:12 crc kubenswrapper[4876]: I0313 07:44:12.904903 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c988b6444-tdqj6"] Mar 13 07:44:12 crc kubenswrapper[4876]: I0313 07:44:12.907320 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" podUID="4a66d689-db59-4913-8e1f-d108bf1fe370" containerName="controller-manager" containerID="cri-o://d3da9ca0e252d4e162916f4ec0c853dbe9a056e1f286a15372d3a05b319a9ddb" gracePeriod=30 Mar 13 07:44:12 crc kubenswrapper[4876]: I0313 07:44:12.997692 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c"] Mar 13 07:44:12 crc kubenswrapper[4876]: I0313 07:44:12.999398 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" podUID="8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" containerName="route-controller-manager" containerID="cri-o://dd88861adab01940a851ebcd8cbefba44db0e5bbadef910dded117593827a74f" gracePeriod=30 Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.042877 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a66d689_db59_4913_8e1f_d108bf1fe370.slice/crio-d3da9ca0e252d4e162916f4ec0c853dbe9a056e1f286a15372d3a05b319a9ddb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a66d689_db59_4913_8e1f_d108bf1fe370.slice/crio-conmon-d3da9ca0e252d4e162916f4ec0c853dbe9a056e1f286a15372d3a05b319a9ddb.scope\": RecentStats: unable to find data in memory cache]" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.497109 4876 generic.go:334] "Generic (PLEG): container finished" podID="8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" containerID="dd88861adab01940a851ebcd8cbefba44db0e5bbadef910dded117593827a74f" exitCode=0 Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.497224 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" event={"ID":"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f","Type":"ContainerDied","Data":"dd88861adab01940a851ebcd8cbefba44db0e5bbadef910dded117593827a74f"} Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.497345 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" event={"ID":"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f","Type":"ContainerDied","Data":"0aeb5f7210246c2de9fc8e4e18f5480a980406b46f16dea10a0b81079fdfebc8"} Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.497361 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0aeb5f7210246c2de9fc8e4e18f5480a980406b46f16dea10a0b81079fdfebc8" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.499374 4876 generic.go:334] "Generic (PLEG): container finished" podID="4a66d689-db59-4913-8e1f-d108bf1fe370" containerID="d3da9ca0e252d4e162916f4ec0c853dbe9a056e1f286a15372d3a05b319a9ddb" exitCode=0 Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.499437 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" event={"ID":"4a66d689-db59-4913-8e1f-d108bf1fe370","Type":"ContainerDied","Data":"d3da9ca0e252d4e162916f4ec0c853dbe9a056e1f286a15372d3a05b319a9ddb"} Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.504963 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.539335 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.547691 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-config\") pod \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.547867 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-serving-cert\") pod \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.548016 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r75sb\" (UniqueName: \"kubernetes.io/projected/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-kube-api-access-r75sb\") pod \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.548195 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-client-ca\") pod \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\" (UID: \"8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.549017 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-config" (OuterVolumeSpecName: "config") pod "8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" (UID: "8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.549032 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-client-ca" (OuterVolumeSpecName: "client-ca") pod "8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" (UID: "8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.555429 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" (UID: "8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.555435 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-kube-api-access-r75sb" (OuterVolumeSpecName: "kube-api-access-r75sb") pod "8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" (UID: "8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f"). InnerVolumeSpecName "kube-api-access-r75sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.649753 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-proxy-ca-bundles\") pod \"4a66d689-db59-4913-8e1f-d108bf1fe370\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.649836 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzr2s\" (UniqueName: \"kubernetes.io/projected/4a66d689-db59-4913-8e1f-d108bf1fe370-kube-api-access-rzr2s\") pod \"4a66d689-db59-4913-8e1f-d108bf1fe370\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.649906 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-config\") pod \"4a66d689-db59-4913-8e1f-d108bf1fe370\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.649965 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a66d689-db59-4913-8e1f-d108bf1fe370-serving-cert\") pod \"4a66d689-db59-4913-8e1f-d108bf1fe370\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.650053 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-client-ca\") pod \"4a66d689-db59-4913-8e1f-d108bf1fe370\" (UID: \"4a66d689-db59-4913-8e1f-d108bf1fe370\") " Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.650382 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.650401 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r75sb\" (UniqueName: \"kubernetes.io/projected/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-kube-api-access-r75sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.650416 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.650428 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.650997 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-config" (OuterVolumeSpecName: "config") pod "4a66d689-db59-4913-8e1f-d108bf1fe370" (UID: "4a66d689-db59-4913-8e1f-d108bf1fe370"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.651017 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-client-ca" (OuterVolumeSpecName: "client-ca") pod "4a66d689-db59-4913-8e1f-d108bf1fe370" (UID: "4a66d689-db59-4913-8e1f-d108bf1fe370"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.651322 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4a66d689-db59-4913-8e1f-d108bf1fe370" (UID: "4a66d689-db59-4913-8e1f-d108bf1fe370"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.653146 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a66d689-db59-4913-8e1f-d108bf1fe370-kube-api-access-rzr2s" (OuterVolumeSpecName: "kube-api-access-rzr2s") pod "4a66d689-db59-4913-8e1f-d108bf1fe370" (UID: "4a66d689-db59-4913-8e1f-d108bf1fe370"). InnerVolumeSpecName "kube-api-access-rzr2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.654866 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a66d689-db59-4913-8e1f-d108bf1fe370-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4a66d689-db59-4913-8e1f-d108bf1fe370" (UID: "4a66d689-db59-4913-8e1f-d108bf1fe370"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.751127 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.751176 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a66d689-db59-4913-8e1f-d108bf1fe370-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.751186 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.751220 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a66d689-db59-4913-8e1f-d108bf1fe370-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.751267 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzr2s\" (UniqueName: \"kubernetes.io/projected/4a66d689-db59-4913-8e1f-d108bf1fe370-kube-api-access-rzr2s\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.989998 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7879bb99b6-zgqhs"] Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990405 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134014ac-2aa1-411c-a5af-492088dee314" containerName="oc" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990427 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="134014ac-2aa1-411c-a5af-492088dee314" containerName="oc" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990450 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerName="extract-content" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990459 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerName="extract-content" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990476 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerName="extract-utilities" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990490 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerName="extract-utilities" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990507 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerName="extract-utilities" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990517 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerName="extract-utilities" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990533 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990541 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990556 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990564 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990577 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerName="extract-content" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990588 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerName="extract-content" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990603 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990613 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990627 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" containerName="route-controller-manager" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990638 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" containerName="route-controller-manager" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990653 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerName="extract-content" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990664 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerName="extract-content" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990679 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a66d689-db59-4913-8e1f-d108bf1fe370" containerName="controller-manager" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990689 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a66d689-db59-4913-8e1f-d108bf1fe370" containerName="controller-manager" Mar 13 07:44:13 crc kubenswrapper[4876]: E0313 07:44:13.990710 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerName="extract-utilities" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990721 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerName="extract-utilities" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990867 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" containerName="route-controller-manager" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990884 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="68fd169c-1d47-4f5b-a528-201a1ea159a1" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990904 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="134014ac-2aa1-411c-a5af-492088dee314" containerName="oc" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990915 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a66d689-db59-4913-8e1f-d108bf1fe370" containerName="controller-manager" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990923 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddbd46d5-05cd-4d71-b085-142b13687f63" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.990931 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4259aab-8c71-4b40-957e-ef4e5ddc1ea1" containerName="registry-server" Mar 13 07:44:13 crc kubenswrapper[4876]: I0313 07:44:13.991560 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.016837 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7879bb99b6-zgqhs"] Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.054596 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-client-ca\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.054655 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-proxy-ca-bundles\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.054691 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb68b\" (UniqueName: \"kubernetes.io/projected/5347e6b6-204e-47b9-a416-01935cb777c5-kube-api-access-zb68b\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.054735 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5347e6b6-204e-47b9-a416-01935cb777c5-serving-cert\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.054773 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-config\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.157458 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-config\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.157693 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-client-ca\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.157783 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-proxy-ca-bundles\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.157905 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb68b\" (UniqueName: \"kubernetes.io/projected/5347e6b6-204e-47b9-a416-01935cb777c5-kube-api-access-zb68b\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.158006 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5347e6b6-204e-47b9-a416-01935cb777c5-serving-cert\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.160037 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-client-ca\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.160302 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-proxy-ca-bundles\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.160627 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5347e6b6-204e-47b9-a416-01935cb777c5-config\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.173932 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5347e6b6-204e-47b9-a416-01935cb777c5-serving-cert\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.196364 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb68b\" (UniqueName: \"kubernetes.io/projected/5347e6b6-204e-47b9-a416-01935cb777c5-kube-api-access-zb68b\") pod \"controller-manager-7879bb99b6-zgqhs\" (UID: \"5347e6b6-204e-47b9-a416-01935cb777c5\") " pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.328796 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.507881 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.508936 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.509087 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c988b6444-tdqj6" event={"ID":"4a66d689-db59-4913-8e1f-d108bf1fe370","Type":"ContainerDied","Data":"9fe12112194182acff6abd45c73b85933bf552778d6b16dae80ec9bfda38d3e9"} Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.511151 4876 scope.go:117] "RemoveContainer" containerID="d3da9ca0e252d4e162916f4ec0c853dbe9a056e1f286a15372d3a05b319a9ddb" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.557465 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c988b6444-tdqj6"] Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.559081 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5c988b6444-tdqj6"] Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.579299 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c"] Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.582730 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6999bfff6-6kw2c"] Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.750853 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7879bb99b6-zgqhs"] Mar 13 07:44:14 crc kubenswrapper[4876]: W0313 07:44:14.759737 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5347e6b6_204e_47b9_a416_01935cb777c5.slice/crio-eabeb85d5132506dbaab5b19442dd839d8166cf2febbbb8be5a58eef411e0b0e WatchSource:0}: Error finding container eabeb85d5132506dbaab5b19442dd839d8166cf2febbbb8be5a58eef411e0b0e: Status 404 returned error can't find the container with id eabeb85d5132506dbaab5b19442dd839d8166cf2febbbb8be5a58eef411e0b0e Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.989956 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd"] Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.991409 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.993967 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.994060 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.999154 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 07:44:14 crc kubenswrapper[4876]: I0313 07:44:14.999816 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.000098 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.001406 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.008391 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd"] Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.049745 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a66d689-db59-4913-8e1f-d108bf1fe370" path="/var/lib/kubelet/pods/4a66d689-db59-4913-8e1f-d108bf1fe370/volumes" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.050799 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f" path="/var/lib/kubelet/pods/8f738ddb-8c69-4dfe-847d-cfaf2ca8bc4f/volumes" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.070631 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6076eba-4629-4b64-834a-9da54288dd13-config\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.070757 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6076eba-4629-4b64-834a-9da54288dd13-serving-cert\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.070805 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg9nb\" (UniqueName: \"kubernetes.io/projected/b6076eba-4629-4b64-834a-9da54288dd13-kube-api-access-jg9nb\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.070964 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6076eba-4629-4b64-834a-9da54288dd13-client-ca\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.119270 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xvfbt"] Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.171689 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6076eba-4629-4b64-834a-9da54288dd13-client-ca\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.172133 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6076eba-4629-4b64-834a-9da54288dd13-config\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.172172 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6076eba-4629-4b64-834a-9da54288dd13-serving-cert\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.172194 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg9nb\" (UniqueName: \"kubernetes.io/projected/b6076eba-4629-4b64-834a-9da54288dd13-kube-api-access-jg9nb\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.173054 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6076eba-4629-4b64-834a-9da54288dd13-client-ca\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.174808 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6076eba-4629-4b64-834a-9da54288dd13-config\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.189134 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6076eba-4629-4b64-834a-9da54288dd13-serving-cert\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.216303 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg9nb\" (UniqueName: \"kubernetes.io/projected/b6076eba-4629-4b64-834a-9da54288dd13-kube-api-access-jg9nb\") pod \"route-controller-manager-7b7b5f9ffd-mzxbd\" (UID: \"b6076eba-4629-4b64-834a-9da54288dd13\") " pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.312301 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.522802 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" event={"ID":"5347e6b6-204e-47b9-a416-01935cb777c5","Type":"ContainerStarted","Data":"22bc2b00f6d3b0495d5c24dfd1fd4b978dbb970c7a3131939f9d76d2d320e582"} Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.523373 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" event={"ID":"5347e6b6-204e-47b9-a416-01935cb777c5","Type":"ContainerStarted","Data":"eabeb85d5132506dbaab5b19442dd839d8166cf2febbbb8be5a58eef411e0b0e"} Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.523394 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.571471 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" podStartSLOduration=3.5714323 podStartE2EDuration="3.5714323s" podCreationTimestamp="2026-03-13 07:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:44:15.570125026 +0000 UTC m=+315.240904008" watchObservedRunningTime="2026-03-13 07:44:15.5714323 +0000 UTC m=+315.242211282" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.578396 4876 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.579300 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.580008 4876 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.580247 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6" gracePeriod=15 Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.580467 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc" gracePeriod=15 Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.580510 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a" gracePeriod=15 Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.580540 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598" gracePeriod=15 Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.580574 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b" gracePeriod=15 Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.598967 4876 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.599775 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.599877 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.599978 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.600054 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.600146 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.600222 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.600335 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.600411 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.600503 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.600580 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.600661 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.600750 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.600831 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.600906 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.600983 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.601063 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.601142 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.601219 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.601467 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.605584 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.605682 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.605767 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.605845 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.605915 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.605980 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.606093 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.606456 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.606557 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.606770 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.613609 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7879bb99b6-zgqhs" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.677025 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.681407 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.681468 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.681506 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.681576 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.681709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.681747 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.681779 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.681799 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.785505 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786194 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786266 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786292 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786317 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786338 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786426 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786456 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786564 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786629 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786662 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786696 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786721 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786749 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786785 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.786815 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: I0313 07:44:15.954273 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:44:15 crc kubenswrapper[4876]: W0313 07:44:15.974904 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-24863c95cdc53e9790409b4ca2c94bbd900247b1e8ecce3405236bb4983432db WatchSource:0}: Error finding container 24863c95cdc53e9790409b4ca2c94bbd900247b1e8ecce3405236bb4983432db: Status 404 returned error can't find the container with id 24863c95cdc53e9790409b4ca2c94bbd900247b1e8ecce3405236bb4983432db Mar 13 07:44:15 crc kubenswrapper[4876]: E0313 07:44:15.979538 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.39:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c56dbf4980999 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:44:15.978563993 +0000 UTC m=+315.649342965,LastTimestamp:2026-03-13 07:44:15.978563993 +0000 UTC m=+315.649342965,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:44:16 crc kubenswrapper[4876]: E0313 07:44:16.026346 4876 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 07:44:16 crc kubenswrapper[4876]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager_b6076eba-4629-4b64-834a-9da54288dd13_0(da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d): error adding pod openshift-route-controller-manager_route-controller-manager-7b7b5f9ffd-mzxbd to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d" Netns:"/var/run/netns/fceed96e-ae21-483a-bf49-aafa499d3e35" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7b7b5f9ffd-mzxbd;K8S_POD_INFRA_CONTAINER_ID=da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d;K8S_POD_UID=b6076eba-4629-4b64-834a-9da54288dd13" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd/b6076eba-4629-4b64-834a-9da54288dd13]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7b7b5f9ffd-mzxbd?timeout=1m0s": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:44:16 crc kubenswrapper[4876]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 07:44:16 crc kubenswrapper[4876]: > Mar 13 07:44:16 crc kubenswrapper[4876]: E0313 07:44:16.026468 4876 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 07:44:16 crc kubenswrapper[4876]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager_b6076eba-4629-4b64-834a-9da54288dd13_0(da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d): error adding pod openshift-route-controller-manager_route-controller-manager-7b7b5f9ffd-mzxbd to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d" Netns:"/var/run/netns/fceed96e-ae21-483a-bf49-aafa499d3e35" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7b7b5f9ffd-mzxbd;K8S_POD_INFRA_CONTAINER_ID=da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d;K8S_POD_UID=b6076eba-4629-4b64-834a-9da54288dd13" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd/b6076eba-4629-4b64-834a-9da54288dd13]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7b7b5f9ffd-mzxbd?timeout=1m0s": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:44:16 crc kubenswrapper[4876]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 07:44:16 crc kubenswrapper[4876]: > pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:16 crc kubenswrapper[4876]: E0313 07:44:16.026493 4876 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 07:44:16 crc kubenswrapper[4876]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager_b6076eba-4629-4b64-834a-9da54288dd13_0(da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d): error adding pod openshift-route-controller-manager_route-controller-manager-7b7b5f9ffd-mzxbd to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d" Netns:"/var/run/netns/fceed96e-ae21-483a-bf49-aafa499d3e35" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7b7b5f9ffd-mzxbd;K8S_POD_INFRA_CONTAINER_ID=da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d;K8S_POD_UID=b6076eba-4629-4b64-834a-9da54288dd13" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd/b6076eba-4629-4b64-834a-9da54288dd13]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7b7b5f9ffd-mzxbd?timeout=1m0s": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:44:16 crc kubenswrapper[4876]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 07:44:16 crc kubenswrapper[4876]: > pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:16 crc kubenswrapper[4876]: E0313 07:44:16.026568 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager(b6076eba-4629-4b64-834a-9da54288dd13)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager(b6076eba-4629-4b64-834a-9da54288dd13)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager_b6076eba-4629-4b64-834a-9da54288dd13_0(da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d): error adding pod openshift-route-controller-manager_route-controller-manager-7b7b5f9ffd-mzxbd to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d\\\" Netns:\\\"/var/run/netns/fceed96e-ae21-483a-bf49-aafa499d3e35\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7b7b5f9ffd-mzxbd;K8S_POD_INFRA_CONTAINER_ID=da217190d303c6e62efc65ae2d15ddac6b8fec8d1dd60f159eaaa2e043c9bc5d;K8S_POD_UID=b6076eba-4629-4b64-834a-9da54288dd13\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd/b6076eba-4629-4b64-834a-9da54288dd13]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7b7b5f9ffd-mzxbd?timeout=1m0s\\\": dial tcp 38.102.83.39:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" podUID="b6076eba-4629-4b64-834a-9da54288dd13" Mar 13 07:44:16 crc kubenswrapper[4876]: E0313 07:44:16.522846 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.39:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c56dbf4980999 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:44:15.978563993 +0000 UTC m=+315.649342965,LastTimestamp:2026-03-13 07:44:15.978563993 +0000 UTC m=+315.649342965,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.534622 4876 generic.go:334] "Generic (PLEG): container finished" podID="de3aea18-c726-431d-97ed-af1a5341ffb5" containerID="87f4ab1a9aaecfbeb6146e65bbd991b5eb6b25ee11b77b665e265cfaf6ea4cfe" exitCode=0 Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.534706 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"de3aea18-c726-431d-97ed-af1a5341ffb5","Type":"ContainerDied","Data":"87f4ab1a9aaecfbeb6146e65bbd991b5eb6b25ee11b77b665e265cfaf6ea4cfe"} Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.537462 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ebe01a9218f14fa2b963c4a3e694ae49107fac9412b157ec4178fe1304bbe3d3"} Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.537497 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"24863c95cdc53e9790409b4ca2c94bbd900247b1e8ecce3405236bb4983432db"} Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.541326 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.543765 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.545088 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc" exitCode=0 Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.545123 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a" exitCode=0 Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.545135 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598" exitCode=0 Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.545144 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b" exitCode=2 Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.545257 4876 scope.go:117] "RemoveContainer" containerID="2aca8add525aa2fa4e43009f7f2c29fc7aa1cc6b195cda1aa68e62b13c223997" Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.545439 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:16 crc kubenswrapper[4876]: I0313 07:44:16.546126 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:17 crc kubenswrapper[4876]: E0313 07:44:17.279683 4876 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 07:44:17 crc kubenswrapper[4876]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager_b6076eba-4629-4b64-834a-9da54288dd13_0(7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b): error adding pod openshift-route-controller-manager_route-controller-manager-7b7b5f9ffd-mzxbd to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b" Netns:"/var/run/netns/86fae193-0976-494d-9845-e06717f4843d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7b7b5f9ffd-mzxbd;K8S_POD_INFRA_CONTAINER_ID=7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b;K8S_POD_UID=b6076eba-4629-4b64-834a-9da54288dd13" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd/b6076eba-4629-4b64-834a-9da54288dd13]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7b7b5f9ffd-mzxbd?timeout=1m0s": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:44:17 crc kubenswrapper[4876]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 07:44:17 crc kubenswrapper[4876]: > Mar 13 07:44:17 crc kubenswrapper[4876]: E0313 07:44:17.280166 4876 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 07:44:17 crc kubenswrapper[4876]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager_b6076eba-4629-4b64-834a-9da54288dd13_0(7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b): error adding pod openshift-route-controller-manager_route-controller-manager-7b7b5f9ffd-mzxbd to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b" Netns:"/var/run/netns/86fae193-0976-494d-9845-e06717f4843d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7b7b5f9ffd-mzxbd;K8S_POD_INFRA_CONTAINER_ID=7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b;K8S_POD_UID=b6076eba-4629-4b64-834a-9da54288dd13" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd/b6076eba-4629-4b64-834a-9da54288dd13]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7b7b5f9ffd-mzxbd?timeout=1m0s": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:44:17 crc kubenswrapper[4876]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 07:44:17 crc kubenswrapper[4876]: > pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:17 crc kubenswrapper[4876]: E0313 07:44:17.280195 4876 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 07:44:17 crc kubenswrapper[4876]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager_b6076eba-4629-4b64-834a-9da54288dd13_0(7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b): error adding pod openshift-route-controller-manager_route-controller-manager-7b7b5f9ffd-mzxbd to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b" Netns:"/var/run/netns/86fae193-0976-494d-9845-e06717f4843d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7b7b5f9ffd-mzxbd;K8S_POD_INFRA_CONTAINER_ID=7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b;K8S_POD_UID=b6076eba-4629-4b64-834a-9da54288dd13" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd/b6076eba-4629-4b64-834a-9da54288dd13]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7b7b5f9ffd-mzxbd?timeout=1m0s": dial tcp 38.102.83.39:6443: connect: connection refused Mar 13 07:44:17 crc kubenswrapper[4876]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 07:44:17 crc kubenswrapper[4876]: > pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:17 crc kubenswrapper[4876]: E0313 07:44:17.280305 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager(b6076eba-4629-4b64-834a-9da54288dd13)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager(b6076eba-4629-4b64-834a-9da54288dd13)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7b7b5f9ffd-mzxbd_openshift-route-controller-manager_b6076eba-4629-4b64-834a-9da54288dd13_0(7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b): error adding pod openshift-route-controller-manager_route-controller-manager-7b7b5f9ffd-mzxbd to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b\\\" Netns:\\\"/var/run/netns/86fae193-0976-494d-9845-e06717f4843d\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7b7b5f9ffd-mzxbd;K8S_POD_INFRA_CONTAINER_ID=7df44b4411ee3535cc576b5d01177f4b82581131c03fddbe1cc1c02dbe120a7b;K8S_POD_UID=b6076eba-4629-4b64-834a-9da54288dd13\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd/b6076eba-4629-4b64-834a-9da54288dd13]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-7b7b5f9ffd-mzxbd in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7b7b5f9ffd-mzxbd?timeout=1m0s\\\": dial tcp 38.102.83.39:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" podUID="b6076eba-4629-4b64-834a-9da54288dd13" Mar 13 07:44:17 crc kubenswrapper[4876]: I0313 07:44:17.556901 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.074950 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.081802 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.082798 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121137 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de3aea18-c726-431d-97ed-af1a5341ffb5-kube-api-access\") pod \"de3aea18-c726-431d-97ed-af1a5341ffb5\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121215 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121285 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-var-lock\") pod \"de3aea18-c726-431d-97ed-af1a5341ffb5\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121309 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121341 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121411 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-kubelet-dir\") pod \"de3aea18-c726-431d-97ed-af1a5341ffb5\" (UID: \"de3aea18-c726-431d-97ed-af1a5341ffb5\") " Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121531 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-var-lock" (OuterVolumeSpecName: "var-lock") pod "de3aea18-c726-431d-97ed-af1a5341ffb5" (UID: "de3aea18-c726-431d-97ed-af1a5341ffb5"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121751 4876 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121793 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "de3aea18-c726-431d-97ed-af1a5341ffb5" (UID: "de3aea18-c726-431d-97ed-af1a5341ffb5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121832 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121857 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.121880 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.128888 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3aea18-c726-431d-97ed-af1a5341ffb5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "de3aea18-c726-431d-97ed-af1a5341ffb5" (UID: "de3aea18-c726-431d-97ed-af1a5341ffb5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.223484 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de3aea18-c726-431d-97ed-af1a5341ffb5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.223578 4876 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.223594 4876 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.223606 4876 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.223619 4876 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/de3aea18-c726-431d-97ed-af1a5341ffb5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.571607 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.572679 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6" exitCode=0 Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.572786 4876 scope.go:117] "RemoveContainer" containerID="3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.572964 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.577491 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"de3aea18-c726-431d-97ed-af1a5341ffb5","Type":"ContainerDied","Data":"832be911592e1c05faf0c48dc92f243f41b4a87d90e21a017b640f2fa0202a3f"} Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.577546 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="832be911592e1c05faf0c48dc92f243f41b4a87d90e21a017b640f2fa0202a3f" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.577613 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.599991 4876 scope.go:117] "RemoveContainer" containerID="de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.618547 4876 scope.go:117] "RemoveContainer" containerID="a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.643314 4876 scope.go:117] "RemoveContainer" containerID="6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.664410 4876 scope.go:117] "RemoveContainer" containerID="593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.689061 4876 scope.go:117] "RemoveContainer" containerID="b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.714486 4876 scope.go:117] "RemoveContainer" containerID="3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc" Mar 13 07:44:18 crc kubenswrapper[4876]: E0313 07:44:18.715310 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\": container with ID starting with 3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc not found: ID does not exist" containerID="3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.715389 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc"} err="failed to get container status \"3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\": rpc error: code = NotFound desc = could not find container \"3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc\": container with ID starting with 3e56468d4c80f4afab4510b239a74a994c72cfd67bbe4d0ec7107349dd2713bc not found: ID does not exist" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.715430 4876 scope.go:117] "RemoveContainer" containerID="de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a" Mar 13 07:44:18 crc kubenswrapper[4876]: E0313 07:44:18.715854 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\": container with ID starting with de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a not found: ID does not exist" containerID="de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.715889 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a"} err="failed to get container status \"de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\": rpc error: code = NotFound desc = could not find container \"de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a\": container with ID starting with de2d10f8fdad25cd427f8edaa9d399bc79f544ada54896a04c8853a25a7e602a not found: ID does not exist" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.715918 4876 scope.go:117] "RemoveContainer" containerID="a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598" Mar 13 07:44:18 crc kubenswrapper[4876]: E0313 07:44:18.716202 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\": container with ID starting with a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598 not found: ID does not exist" containerID="a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.716272 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598"} err="failed to get container status \"a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\": rpc error: code = NotFound desc = could not find container \"a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598\": container with ID starting with a7910d99037459da07b595044741dc8b7b1dbaaa0d845807f422fd5fe8f7d598 not found: ID does not exist" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.716332 4876 scope.go:117] "RemoveContainer" containerID="6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b" Mar 13 07:44:18 crc kubenswrapper[4876]: E0313 07:44:18.716620 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\": container with ID starting with 6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b not found: ID does not exist" containerID="6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.716648 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b"} err="failed to get container status \"6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\": rpc error: code = NotFound desc = could not find container \"6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b\": container with ID starting with 6be97ded9799affa03060337c3cf24749fc79d79706bd5a1718956f43e2e672b not found: ID does not exist" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.716666 4876 scope.go:117] "RemoveContainer" containerID="593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6" Mar 13 07:44:18 crc kubenswrapper[4876]: E0313 07:44:18.716884 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\": container with ID starting with 593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6 not found: ID does not exist" containerID="593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.716907 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6"} err="failed to get container status \"593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\": rpc error: code = NotFound desc = could not find container \"593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6\": container with ID starting with 593432c787c1bc54a50cb678ec720027345fbecbfb43627c6720c8f253cae9c6 not found: ID does not exist" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.716921 4876 scope.go:117] "RemoveContainer" containerID="b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a" Mar 13 07:44:18 crc kubenswrapper[4876]: E0313 07:44:18.717100 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\": container with ID starting with b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a not found: ID does not exist" containerID="b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a" Mar 13 07:44:18 crc kubenswrapper[4876]: I0313 07:44:18.717123 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a"} err="failed to get container status \"b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\": rpc error: code = NotFound desc = could not find container \"b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a\": container with ID starting with b3065ab566f02d6315dc8e51c3c81e4cd046aec8e408d88cfb00cf7fdeacfd9a not found: ID does not exist" Mar 13 07:44:19 crc kubenswrapper[4876]: I0313 07:44:19.045763 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 13 07:44:19 crc kubenswrapper[4876]: E0313 07:44:19.134029 4876 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.39:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" volumeName="registry-storage" Mar 13 07:44:20 crc kubenswrapper[4876]: I0313 07:44:20.657121 4876 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:20 crc kubenswrapper[4876]: I0313 07:44:20.657885 4876 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:20 crc kubenswrapper[4876]: I0313 07:44:20.658326 4876 status_manager.go:851] "Failed to get status for pod" podUID="de3aea18-c726-431d-97ed-af1a5341ffb5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:21 crc kubenswrapper[4876]: I0313 07:44:21.040729 4876 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:21 crc kubenswrapper[4876]: I0313 07:44:21.041202 4876 status_manager.go:851] "Failed to get status for pod" podUID="de3aea18-c726-431d-97ed-af1a5341ffb5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:22 crc kubenswrapper[4876]: E0313 07:44:22.865142 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:22 crc kubenswrapper[4876]: E0313 07:44:22.867437 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:22 crc kubenswrapper[4876]: E0313 07:44:22.868726 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:22 crc kubenswrapper[4876]: E0313 07:44:22.869744 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:22 crc kubenswrapper[4876]: E0313 07:44:22.870518 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:22 crc kubenswrapper[4876]: I0313 07:44:22.870608 4876 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 13 07:44:22 crc kubenswrapper[4876]: E0313 07:44:22.871201 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="200ms" Mar 13 07:44:23 crc kubenswrapper[4876]: E0313 07:44:23.072522 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="400ms" Mar 13 07:44:23 crc kubenswrapper[4876]: E0313 07:44:23.473290 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="800ms" Mar 13 07:44:24 crc kubenswrapper[4876]: E0313 07:44:24.275436 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="1.6s" Mar 13 07:44:25 crc kubenswrapper[4876]: E0313 07:44:25.876905 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="3.2s" Mar 13 07:44:26 crc kubenswrapper[4876]: E0313 07:44:26.524707 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.39:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c56dbf4980999 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 07:44:15.978563993 +0000 UTC m=+315.649342965,LastTimestamp:2026-03-13 07:44:15.978563993 +0000 UTC m=+315.649342965,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.038044 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.040550 4876 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.041382 4876 status_manager.go:851] "Failed to get status for pod" podUID="de3aea18-c726-431d-97ed-af1a5341ffb5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.065558 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.065607 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:27 crc kubenswrapper[4876]: E0313 07:44:27.066438 4876 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.067183 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:27 crc kubenswrapper[4876]: W0313 07:44:27.097622 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-f54574b5883b28085a10d19df048314a17b2107f78641e322e7387dac5e63693 WatchSource:0}: Error finding container f54574b5883b28085a10d19df048314a17b2107f78641e322e7387dac5e63693: Status 404 returned error can't find the container with id f54574b5883b28085a10d19df048314a17b2107f78641e322e7387dac5e63693 Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.641355 4876 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="cdcf99fc615f328a4b729ad9a1c98152afd1db9cb759601e4b68014ed8a0f03c" exitCode=0 Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.641442 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"cdcf99fc615f328a4b729ad9a1c98152afd1db9cb759601e4b68014ed8a0f03c"} Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.641845 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f54574b5883b28085a10d19df048314a17b2107f78641e322e7387dac5e63693"} Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.642162 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.642177 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.642993 4876 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:27 crc kubenswrapper[4876]: E0313 07:44:27.643113 4876 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:27 crc kubenswrapper[4876]: I0313 07:44:27.643465 4876 status_manager.go:851] "Failed to get status for pod" podUID="de3aea18-c726-431d-97ed-af1a5341ffb5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Mar 13 07:44:28 crc kubenswrapper[4876]: I0313 07:44:28.667556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"05ec88bf6cbb0e51ed55df0482e152593f9bd93af63e7a7e8c1611be764410c4"} Mar 13 07:44:28 crc kubenswrapper[4876]: I0313 07:44:28.667623 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"914ecfb14818493ec75f3bc61f1cf5148158396c0c40f0c7aa56c4d524872783"} Mar 13 07:44:28 crc kubenswrapper[4876]: I0313 07:44:28.667634 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6c8a7a52360ec7c6311f7c55d837b1927bea76e37bacf4d0ecce9c6994c7cba1"} Mar 13 07:44:28 crc kubenswrapper[4876]: I0313 07:44:28.667644 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"14aea3e09d01eabe3899dd9ff654a3c9d1f08895e20e750889f8728c96eac89d"} Mar 13 07:44:29 crc kubenswrapper[4876]: I0313 07:44:29.677156 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7006c41d0652103a5e1201e66bdc1df59e4084e5b9231360abed5669293ca9e3"} Mar 13 07:44:29 crc kubenswrapper[4876]: I0313 07:44:29.677754 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:29 crc kubenswrapper[4876]: I0313 07:44:29.677635 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:29 crc kubenswrapper[4876]: I0313 07:44:29.677790 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:30 crc kubenswrapper[4876]: I0313 07:44:30.034622 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:30 crc kubenswrapper[4876]: I0313 07:44:30.035177 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:30 crc kubenswrapper[4876]: I0313 07:44:30.688713 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 07:44:30 crc kubenswrapper[4876]: I0313 07:44:30.689355 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 07:44:30 crc kubenswrapper[4876]: I0313 07:44:30.689398 4876 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e" exitCode=1 Mar 13 07:44:30 crc kubenswrapper[4876]: I0313 07:44:30.689440 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e"} Mar 13 07:44:30 crc kubenswrapper[4876]: I0313 07:44:30.690053 4876 scope.go:117] "RemoveContainer" containerID="d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e" Mar 13 07:44:31 crc kubenswrapper[4876]: I0313 07:44:31.114974 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:44:31 crc kubenswrapper[4876]: I0313 07:44:31.706035 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 07:44:31 crc kubenswrapper[4876]: I0313 07:44:31.707663 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 07:44:31 crc kubenswrapper[4876]: I0313 07:44:31.707749 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5a600ee1e9478878cde275c1e7b50f30ab58443ecfc74342eee07ebd3ff67580"} Mar 13 07:44:32 crc kubenswrapper[4876]: I0313 07:44:32.072032 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:32 crc kubenswrapper[4876]: I0313 07:44:32.072997 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:32 crc kubenswrapper[4876]: I0313 07:44:32.080399 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:34 crc kubenswrapper[4876]: I0313 07:44:34.691553 4876 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:34 crc kubenswrapper[4876]: I0313 07:44:34.729470 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:34 crc kubenswrapper[4876]: I0313 07:44:34.729862 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:34 crc kubenswrapper[4876]: I0313 07:44:34.734303 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:34 crc kubenswrapper[4876]: I0313 07:44:34.779732 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="52903cb1-4be2-400f-b0ed-25aa82d27957" Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.656415 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.657181 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.657289 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.738912 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" event={"ID":"b6076eba-4629-4b64-834a-9da54288dd13","Type":"ContainerStarted","Data":"68a68b262ae91de5fd79e0264fa2bc1750fec3e3adfded6d5ac6a4c55806926d"} Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.739002 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" event={"ID":"b6076eba-4629-4b64-834a-9da54288dd13","Type":"ContainerStarted","Data":"60bd883acb08e3c1179f36e232d90a3ee6ec054bc3a4e56646d4570e29f160bf"} Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.739298 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.739333 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2acda0f0-90e3-4168-9754-abd0bc4bf37b" Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.739423 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:35 crc kubenswrapper[4876]: I0313 07:44:35.743842 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="52903cb1-4be2-400f-b0ed-25aa82d27957" Mar 13 07:44:36 crc kubenswrapper[4876]: I0313 07:44:36.739571 4876 patch_prober.go:28] interesting pod/route-controller-manager-7b7b5f9ffd-mzxbd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:44:36 crc kubenswrapper[4876]: I0313 07:44:36.739695 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" podUID="b6076eba-4629-4b64-834a-9da54288dd13" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 07:44:37 crc kubenswrapper[4876]: I0313 07:44:37.746229 4876 patch_prober.go:28] interesting pod/route-controller-manager-7b7b5f9ffd-mzxbd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:44:37 crc kubenswrapper[4876]: I0313 07:44:37.746952 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" podUID="b6076eba-4629-4b64-834a-9da54288dd13" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 07:44:40 crc kubenswrapper[4876]: I0313 07:44:40.632862 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" podUID="6b92045d-38c7-4dff-be0a-78449d870bcb" containerName="oauth-openshift" containerID="cri-o://6902aef66af563dcddb390d4ca47a333014cfe448070e37021e928c7d4e9e200" gracePeriod=15 Mar 13 07:44:40 crc kubenswrapper[4876]: I0313 07:44:40.788617 4876 generic.go:334] "Generic (PLEG): container finished" podID="6b92045d-38c7-4dff-be0a-78449d870bcb" containerID="6902aef66af563dcddb390d4ca47a333014cfe448070e37021e928c7d4e9e200" exitCode=0 Mar 13 07:44:40 crc kubenswrapper[4876]: I0313 07:44:40.788706 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" event={"ID":"6b92045d-38c7-4dff-be0a-78449d870bcb","Type":"ContainerDied","Data":"6902aef66af563dcddb390d4ca47a333014cfe448070e37021e928c7d4e9e200"} Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.108745 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.114665 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278001 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-idp-0-file-data\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278071 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf48g\" (UniqueName: \"kubernetes.io/projected/6b92045d-38c7-4dff-be0a-78449d870bcb-kube-api-access-mf48g\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278143 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-trusted-ca-bundle\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278212 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-service-ca\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278331 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-router-certs\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278374 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-error\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278407 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-session\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278454 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-login\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278504 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-policies\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278583 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-serving-cert\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278631 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-cliconfig\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278681 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-provider-selection\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278725 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-dir\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.278776 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-ocp-branding-template\") pod \"6b92045d-38c7-4dff-be0a-78449d870bcb\" (UID: \"6b92045d-38c7-4dff-be0a-78449d870bcb\") " Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.279483 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.279523 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.279564 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.279595 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.280374 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.286993 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.287419 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.287675 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.288440 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.288856 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.291333 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b92045d-38c7-4dff-be0a-78449d870bcb-kube-api-access-mf48g" (OuterVolumeSpecName: "kube-api-access-mf48g") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "kube-api-access-mf48g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.291408 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.296829 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.297027 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "6b92045d-38c7-4dff-be0a-78449d870bcb" (UID: "6b92045d-38c7-4dff-be0a-78449d870bcb"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380538 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380582 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380593 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380605 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380619 4876 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380629 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380639 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380651 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380663 4876 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6b92045d-38c7-4dff-be0a-78449d870bcb-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380673 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380687 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380699 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf48g\" (UniqueName: \"kubernetes.io/projected/6b92045d-38c7-4dff-be0a-78449d870bcb-kube-api-access-mf48g\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380709 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.380718 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6b92045d-38c7-4dff-be0a-78449d870bcb-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.629330 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.796160 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" event={"ID":"6b92045d-38c7-4dff-be0a-78449d870bcb","Type":"ContainerDied","Data":"5949fb0f1aa6fff520cf6b58fce43527d8f22fb89616ab9754b9ee529f10d530"} Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.796218 4876 scope.go:117] "RemoveContainer" containerID="6902aef66af563dcddb390d4ca47a333014cfe448070e37021e928c7d4e9e200" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.796232 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-xvfbt" Mar 13 07:44:41 crc kubenswrapper[4876]: I0313 07:44:41.957364 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 07:44:42 crc kubenswrapper[4876]: I0313 07:44:42.910378 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 07:44:43 crc kubenswrapper[4876]: I0313 07:44:43.317800 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 07:44:43 crc kubenswrapper[4876]: I0313 07:44:43.974717 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 07:44:44 crc kubenswrapper[4876]: I0313 07:44:44.172921 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 07:44:44 crc kubenswrapper[4876]: I0313 07:44:44.195629 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 07:44:44 crc kubenswrapper[4876]: I0313 07:44:44.629545 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 07:44:45 crc kubenswrapper[4876]: I0313 07:44:45.145410 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 07:44:45 crc kubenswrapper[4876]: I0313 07:44:45.654103 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 07:44:45 crc kubenswrapper[4876]: I0313 07:44:45.657304 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 07:44:45 crc kubenswrapper[4876]: I0313 07:44:45.657425 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 07:44:45 crc kubenswrapper[4876]: I0313 07:44:45.751599 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 07:44:45 crc kubenswrapper[4876]: I0313 07:44:45.814563 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 07:44:45 crc kubenswrapper[4876]: I0313 07:44:45.995191 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 07:44:46 crc kubenswrapper[4876]: I0313 07:44:46.313007 4876 patch_prober.go:28] interesting pod/route-controller-manager-7b7b5f9ffd-mzxbd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 07:44:46 crc kubenswrapper[4876]: I0313 07:44:46.313147 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" podUID="b6076eba-4629-4b64-834a-9da54288dd13" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 07:44:46 crc kubenswrapper[4876]: I0313 07:44:46.436116 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 07:44:46 crc kubenswrapper[4876]: I0313 07:44:46.588385 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 07:44:46 crc kubenswrapper[4876]: I0313 07:44:46.764374 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.049242 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.167508 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.200866 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.222609 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.282955 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.297544 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.560889 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.726289 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.759042 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.911068 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.950322 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 07:44:47 crc kubenswrapper[4876]: I0313 07:44:47.958973 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 07:44:48 crc kubenswrapper[4876]: I0313 07:44:48.318322 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 07:44:48 crc kubenswrapper[4876]: I0313 07:44:48.420433 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 07:44:48 crc kubenswrapper[4876]: I0313 07:44:48.577745 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 07:44:48 crc kubenswrapper[4876]: I0313 07:44:48.661396 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 07:44:48 crc kubenswrapper[4876]: I0313 07:44:48.772466 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 07:44:48 crc kubenswrapper[4876]: I0313 07:44:48.824451 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.059201 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.145609 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.149431 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.294642 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.300054 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.342004 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.396722 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.400957 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.574764 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.648018 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.705499 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.741312 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.768782 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.815544 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.854710 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.899771 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.913079 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.944159 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.971576 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 07:44:49 crc kubenswrapper[4876]: I0313 07:44:49.996064 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.017656 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.043022 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.062326 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.067338 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.340424 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.441182 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.503102 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.589324 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.598687 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.688075 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.718789 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.833384 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 07:44:50 crc kubenswrapper[4876]: I0313 07:44:50.952708 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.160917 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.201664 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.294138 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.314178 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.483996 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.648720 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.665809 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.682554 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.684900 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.722654 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.723602 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.750371 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.850959 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.866887 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 07:44:51 crc kubenswrapper[4876]: I0313 07:44:51.934050 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.016086 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.021535 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.066947 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.086702 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.181838 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.215166 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.246513 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.301763 4876 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.302783 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" podStartSLOduration=39.302754535 podStartE2EDuration="39.302754535s" podCreationTimestamp="2026-03-13 07:44:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:44:35.762304934 +0000 UTC m=+335.433083916" watchObservedRunningTime="2026-03-13 07:44:52.302754535 +0000 UTC m=+351.973533517" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.304920 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=37.304911118 podStartE2EDuration="37.304911118s" podCreationTimestamp="2026-03-13 07:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:44:34.721056192 +0000 UTC m=+334.391835184" watchObservedRunningTime="2026-03-13 07:44:52.304911118 +0000 UTC m=+351.975690100" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.307291 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xvfbt","openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.307368 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.307398 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd"] Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.313319 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7b7b5f9ffd-mzxbd" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.328999 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.328967194 podStartE2EDuration="18.328967194s" podCreationTimestamp="2026-03-13 07:44:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:44:52.323702633 +0000 UTC m=+351.994481615" watchObservedRunningTime="2026-03-13 07:44:52.328967194 +0000 UTC m=+351.999746186" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.337417 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.370762 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.564063 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.629464 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.704678 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.740419 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.890289 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.901065 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 07:44:52 crc kubenswrapper[4876]: I0313 07:44:52.966876 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.044486 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b92045d-38c7-4dff-be0a-78449d870bcb" path="/var/lib/kubelet/pods/6b92045d-38c7-4dff-be0a-78449d870bcb/volumes" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.102635 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.149270 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.179529 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.184792 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.243646 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.259635 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.285481 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.348364 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.377787 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.580369 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.626371 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.741481 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.747715 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.816055 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.858606 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.958411 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 07:44:53 crc kubenswrapper[4876]: I0313 07:44:53.973788 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.004266 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.112653 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.157919 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.327316 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.464633 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.497941 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.504028 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.657692 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.703103 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.715816 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.902023 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.931263 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 07:44:54 crc kubenswrapper[4876]: I0313 07:44:54.942289 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.036629 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.134786 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.151642 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.154109 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.260358 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.268448 4876 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.306644 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.435273 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.447649 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.487519 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.503698 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.557267 4876 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.658306 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.658397 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.658467 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.659146 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"5a600ee1e9478878cde275c1e7b50f30ab58443ecfc74342eee07ebd3ff67580"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.659298 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://5a600ee1e9478878cde275c1e7b50f30ab58443ecfc74342eee07ebd3ff67580" gracePeriod=30 Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.805215 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.861862 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.887260 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.889732 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.934388 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.934407 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 07:44:55 crc kubenswrapper[4876]: I0313 07:44:55.966412 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.088408 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.108504 4876 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.231419 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.254414 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.324932 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.365336 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.391804 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.471363 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.485589 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.500921 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.527287 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.531520 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.610003 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.613796 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.665785 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.723944 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.733160 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.751975 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.760286 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.896262 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.946592 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.951190 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.951252 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.956573 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.959021 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.973129 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 07:44:56 crc kubenswrapper[4876]: I0313 07:44:56.998014 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.016814 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.069637 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb"] Mar 13 07:44:57 crc kubenswrapper[4876]: E0313 07:44:57.070007 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b92045d-38c7-4dff-be0a-78449d870bcb" containerName="oauth-openshift" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.070027 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b92045d-38c7-4dff-be0a-78449d870bcb" containerName="oauth-openshift" Mar 13 07:44:57 crc kubenswrapper[4876]: E0313 07:44:57.070042 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3aea18-c726-431d-97ed-af1a5341ffb5" containerName="installer" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.070051 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3aea18-c726-431d-97ed-af1a5341ffb5" containerName="installer" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.070221 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3aea18-c726-431d-97ed-af1a5341ffb5" containerName="installer" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.070260 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b92045d-38c7-4dff-be0a-78449d870bcb" containerName="oauth-openshift" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.070852 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.074076 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.074555 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.074869 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075223 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075477 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075604 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075718 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075763 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075794 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-service-ca\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075834 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075872 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ea62db5-30f6-44dd-aad9-cea501919ece-audit-dir\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075933 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-login\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075962 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-router-certs\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.075996 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-error\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.076039 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-audit-policies\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.076064 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.076099 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.076128 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-serving-cert\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.076150 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-session\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.076179 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwd7m\" (UniqueName: \"kubernetes.io/projected/5ea62db5-30f6-44dd-aad9-cea501919ece-kube-api-access-jwd7m\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.078586 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.080383 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.080605 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.080745 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.080910 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.081214 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.081258 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.083975 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.087838 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb"] Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.094272 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.097046 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.098570 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.109448 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176697 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-login\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176755 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-router-certs\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176789 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-error\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176820 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-audit-policies\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176838 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176859 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-serving-cert\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176878 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176901 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-session\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176924 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwd7m\" (UniqueName: \"kubernetes.io/projected/5ea62db5-30f6-44dd-aad9-cea501919ece-kube-api-access-jwd7m\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176953 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176979 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.176995 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-service-ca\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.177017 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.177041 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ea62db5-30f6-44dd-aad9-cea501919ece-audit-dir\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.177121 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ea62db5-30f6-44dd-aad9-cea501919ece-audit-dir\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.179590 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.179723 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-service-ca\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.179899 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-audit-policies\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.180008 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.184636 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.184690 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-session\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.186334 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.186791 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-login\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.187020 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-user-template-error\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.188022 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-router-certs\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.189039 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-serving-cert\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.207850 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5ea62db5-30f6-44dd-aad9-cea501919ece-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.211605 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwd7m\" (UniqueName: \"kubernetes.io/projected/5ea62db5-30f6-44dd-aad9-cea501919ece-kube-api-access-jwd7m\") pod \"oauth-openshift-77df6bdc9c-qfjsb\" (UID: \"5ea62db5-30f6-44dd-aad9-cea501919ece\") " pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.285296 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.385415 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.386542 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.393331 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.478733 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.754351 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 07:44:57 crc kubenswrapper[4876]: I0313 07:44:57.783652 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.065786 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.163845 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.269567 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb"] Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.285634 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.297807 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.327161 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.366358 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.378779 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.415254 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.486595 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.502753 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.635655 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.776721 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.782097 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.840566 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.932171 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" event={"ID":"5ea62db5-30f6-44dd-aad9-cea501919ece","Type":"ContainerStarted","Data":"79ab8a9013806321c1e3bf0ba3e01a74523e52419fed66861864fc05606da755"} Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.932298 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" event={"ID":"5ea62db5-30f6-44dd-aad9-cea501919ece","Type":"ContainerStarted","Data":"1f52ca8128a73a6423a21b75619eabff93ff9b589d8aa2d082be555670fb9605"} Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.932911 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.957003 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" podStartSLOduration=43.956982562 podStartE2EDuration="43.956982562s" podCreationTimestamp="2026-03-13 07:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:44:58.955173348 +0000 UTC m=+358.625952340" watchObservedRunningTime="2026-03-13 07:44:58.956982562 +0000 UTC m=+358.627761544" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.958191 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-77df6bdc9c-qfjsb" Mar 13 07:44:58 crc kubenswrapper[4876]: I0313 07:44:58.959041 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.047206 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.080782 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.235712 4876 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.370560 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.497624 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.508121 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.643579 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.923624 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 07:44:59 crc kubenswrapper[4876]: I0313 07:44:59.944374 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 07:45:00 crc kubenswrapper[4876]: I0313 07:45:00.061791 4876 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 07:45:00 crc kubenswrapper[4876]: I0313 07:45:00.507352 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 07:45:00 crc kubenswrapper[4876]: I0313 07:45:00.649812 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 07:45:00 crc kubenswrapper[4876]: I0313 07:45:00.670193 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 07:45:00 crc kubenswrapper[4876]: I0313 07:45:00.677045 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 07:45:00 crc kubenswrapper[4876]: I0313 07:45:00.727306 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 07:45:00 crc kubenswrapper[4876]: I0313 07:45:00.768088 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 07:45:00 crc kubenswrapper[4876]: I0313 07:45:00.890866 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.077905 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.112985 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.244030 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.416766 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.434601 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.498956 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.631170 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.721184 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.740319 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.752830 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 07:45:01 crc kubenswrapper[4876]: I0313 07:45:01.910715 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.066219 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.176964 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.288485 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.353177 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.365624 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.409108 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.711982 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.758940 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 07:45:02 crc kubenswrapper[4876]: I0313 07:45:02.871410 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 07:45:03 crc kubenswrapper[4876]: I0313 07:45:03.883380 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 07:45:08 crc kubenswrapper[4876]: I0313 07:45:08.490531 4876 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 07:45:08 crc kubenswrapper[4876]: I0313 07:45:08.491726 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://ebe01a9218f14fa2b963c4a3e694ae49107fac9412b157ec4178fe1304bbe3d3" gracePeriod=5 Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.045463 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.046138 4876 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="ebe01a9218f14fa2b963c4a3e694ae49107fac9412b157ec4178fe1304bbe3d3" exitCode=137 Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.046215 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24863c95cdc53e9790409b4ca2c94bbd900247b1e8ecce3405236bb4983432db" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.071023 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.071117 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.153852 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.153911 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154002 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154043 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154067 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154143 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154176 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154257 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154078 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154633 4876 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154651 4876 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154663 4876 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.154672 4876 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.169947 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:45:14 crc kubenswrapper[4876]: I0313 07:45:14.256942 4876 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 07:45:15 crc kubenswrapper[4876]: I0313 07:45:15.045679 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 13 07:45:15 crc kubenswrapper[4876]: I0313 07:45:15.046476 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 13 07:45:15 crc kubenswrapper[4876]: I0313 07:45:15.051481 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 07:45:15 crc kubenswrapper[4876]: I0313 07:45:15.064536 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 07:45:15 crc kubenswrapper[4876]: I0313 07:45:15.064606 4876 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="19ea077a-19fd-4c38-8520-e128045d6b2c" Mar 13 07:45:15 crc kubenswrapper[4876]: I0313 07:45:15.073018 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 07:45:15 crc kubenswrapper[4876]: I0313 07:45:15.073079 4876 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="19ea077a-19fd-4c38-8520-e128045d6b2c" Mar 13 07:45:26 crc kubenswrapper[4876]: I0313 07:45:26.132177 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 13 07:45:26 crc kubenswrapper[4876]: I0313 07:45:26.134845 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 07:45:26 crc kubenswrapper[4876]: I0313 07:45:26.135467 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 07:45:26 crc kubenswrapper[4876]: I0313 07:45:26.135521 4876 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="5a600ee1e9478878cde275c1e7b50f30ab58443ecfc74342eee07ebd3ff67580" exitCode=137 Mar 13 07:45:26 crc kubenswrapper[4876]: I0313 07:45:26.135554 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"5a600ee1e9478878cde275c1e7b50f30ab58443ecfc74342eee07ebd3ff67580"} Mar 13 07:45:26 crc kubenswrapper[4876]: I0313 07:45:26.135593 4876 scope.go:117] "RemoveContainer" containerID="d55d80ed5d522e71cf3c39c3fbc573ed4bad5816b4e9a4914e0648826414803e" Mar 13 07:45:27 crc kubenswrapper[4876]: I0313 07:45:27.143424 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 13 07:45:27 crc kubenswrapper[4876]: I0313 07:45:27.144683 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 07:45:27 crc kubenswrapper[4876]: I0313 07:45:27.144734 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d7c22340104adae67e374392e7f5480eb09f66d559855b0652b05877cd98e477"} Mar 13 07:45:31 crc kubenswrapper[4876]: I0313 07:45:31.114701 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:45:35 crc kubenswrapper[4876]: I0313 07:45:35.657101 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:45:35 crc kubenswrapper[4876]: I0313 07:45:35.663384 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:45:36 crc kubenswrapper[4876]: I0313 07:45:36.302391 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 07:45:44 crc kubenswrapper[4876]: I0313 07:45:44.954544 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6"] Mar 13 07:45:44 crc kubenswrapper[4876]: E0313 07:45:44.955478 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 07:45:44 crc kubenswrapper[4876]: I0313 07:45:44.955493 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 07:45:44 crc kubenswrapper[4876]: I0313 07:45:44.955603 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 07:45:44 crc kubenswrapper[4876]: I0313 07:45:44.956054 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:44 crc kubenswrapper[4876]: I0313 07:45:44.958577 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 07:45:44 crc kubenswrapper[4876]: I0313 07:45:44.958627 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 07:45:44 crc kubenswrapper[4876]: I0313 07:45:44.967095 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6"] Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.027409 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8c6j\" (UniqueName: \"kubernetes.io/projected/884d6b32-5128-4103-a9b6-82fb23d792ce-kube-api-access-q8c6j\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.027496 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884d6b32-5128-4103-a9b6-82fb23d792ce-config-volume\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.027580 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884d6b32-5128-4103-a9b6-82fb23d792ce-secret-volume\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.129278 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884d6b32-5128-4103-a9b6-82fb23d792ce-config-volume\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.129395 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884d6b32-5128-4103-a9b6-82fb23d792ce-secret-volume\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.129444 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8c6j\" (UniqueName: \"kubernetes.io/projected/884d6b32-5128-4103-a9b6-82fb23d792ce-kube-api-access-q8c6j\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.130556 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884d6b32-5128-4103-a9b6-82fb23d792ce-config-volume\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.142672 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884d6b32-5128-4103-a9b6-82fb23d792ce-secret-volume\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.153900 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8c6j\" (UniqueName: \"kubernetes.io/projected/884d6b32-5128-4103-a9b6-82fb23d792ce-kube-api-access-q8c6j\") pod \"collect-profiles-29556465-pjcv6\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.276247 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:45 crc kubenswrapper[4876]: I0313 07:45:45.731730 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6"] Mar 13 07:45:45 crc kubenswrapper[4876]: W0313 07:45:45.738164 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod884d6b32_5128_4103_a9b6_82fb23d792ce.slice/crio-f2b932db154d43d075bdfcca16bd8da8951275090075573f6c5eb1afe8735a35 WatchSource:0}: Error finding container f2b932db154d43d075bdfcca16bd8da8951275090075573f6c5eb1afe8735a35: Status 404 returned error can't find the container with id f2b932db154d43d075bdfcca16bd8da8951275090075573f6c5eb1afe8735a35 Mar 13 07:45:46 crc kubenswrapper[4876]: I0313 07:45:46.391823 4876 generic.go:334] "Generic (PLEG): container finished" podID="884d6b32-5128-4103-a9b6-82fb23d792ce" containerID="edef007bd0bcea1e3c47cd9810ded713471e6f3dd57dbd7814a0aa908b17dc62" exitCode=0 Mar 13 07:45:46 crc kubenswrapper[4876]: I0313 07:45:46.391915 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" event={"ID":"884d6b32-5128-4103-a9b6-82fb23d792ce","Type":"ContainerDied","Data":"edef007bd0bcea1e3c47cd9810ded713471e6f3dd57dbd7814a0aa908b17dc62"} Mar 13 07:45:46 crc kubenswrapper[4876]: I0313 07:45:46.392178 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" event={"ID":"884d6b32-5128-4103-a9b6-82fb23d792ce","Type":"ContainerStarted","Data":"f2b932db154d43d075bdfcca16bd8da8951275090075573f6c5eb1afe8735a35"} Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.721339 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.868087 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8c6j\" (UniqueName: \"kubernetes.io/projected/884d6b32-5128-4103-a9b6-82fb23d792ce-kube-api-access-q8c6j\") pod \"884d6b32-5128-4103-a9b6-82fb23d792ce\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.868228 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884d6b32-5128-4103-a9b6-82fb23d792ce-secret-volume\") pod \"884d6b32-5128-4103-a9b6-82fb23d792ce\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.868437 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884d6b32-5128-4103-a9b6-82fb23d792ce-config-volume\") pod \"884d6b32-5128-4103-a9b6-82fb23d792ce\" (UID: \"884d6b32-5128-4103-a9b6-82fb23d792ce\") " Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.869085 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/884d6b32-5128-4103-a9b6-82fb23d792ce-config-volume" (OuterVolumeSpecName: "config-volume") pod "884d6b32-5128-4103-a9b6-82fb23d792ce" (UID: "884d6b32-5128-4103-a9b6-82fb23d792ce"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.875854 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/884d6b32-5128-4103-a9b6-82fb23d792ce-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "884d6b32-5128-4103-a9b6-82fb23d792ce" (UID: "884d6b32-5128-4103-a9b6-82fb23d792ce"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.877505 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/884d6b32-5128-4103-a9b6-82fb23d792ce-kube-api-access-q8c6j" (OuterVolumeSpecName: "kube-api-access-q8c6j") pod "884d6b32-5128-4103-a9b6-82fb23d792ce" (UID: "884d6b32-5128-4103-a9b6-82fb23d792ce"). InnerVolumeSpecName "kube-api-access-q8c6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.971140 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8c6j\" (UniqueName: \"kubernetes.io/projected/884d6b32-5128-4103-a9b6-82fb23d792ce-kube-api-access-q8c6j\") on node \"crc\" DevicePath \"\"" Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.971213 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884d6b32-5128-4103-a9b6-82fb23d792ce-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 07:45:47 crc kubenswrapper[4876]: I0313 07:45:47.971291 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884d6b32-5128-4103-a9b6-82fb23d792ce-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 07:45:48 crc kubenswrapper[4876]: I0313 07:45:48.406721 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" event={"ID":"884d6b32-5128-4103-a9b6-82fb23d792ce","Type":"ContainerDied","Data":"f2b932db154d43d075bdfcca16bd8da8951275090075573f6c5eb1afe8735a35"} Mar 13 07:45:48 crc kubenswrapper[4876]: I0313 07:45:48.406809 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2b932db154d43d075bdfcca16bd8da8951275090075573f6c5eb1afe8735a35" Mar 13 07:45:48 crc kubenswrapper[4876]: I0313 07:45:48.406848 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.145098 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556466-zhrhb"] Mar 13 07:46:00 crc kubenswrapper[4876]: E0313 07:46:00.145912 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884d6b32-5128-4103-a9b6-82fb23d792ce" containerName="collect-profiles" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.145927 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="884d6b32-5128-4103-a9b6-82fb23d792ce" containerName="collect-profiles" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.146042 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="884d6b32-5128-4103-a9b6-82fb23d792ce" containerName="collect-profiles" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.146541 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556466-zhrhb" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.149880 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.150890 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.150899 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.156203 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556466-zhrhb"] Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.265942 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fprvl\" (UniqueName: \"kubernetes.io/projected/5ace24c1-074f-4c5b-8148-5bb46a790a3e-kube-api-access-fprvl\") pod \"auto-csr-approver-29556466-zhrhb\" (UID: \"5ace24c1-074f-4c5b-8148-5bb46a790a3e\") " pod="openshift-infra/auto-csr-approver-29556466-zhrhb" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.368325 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fprvl\" (UniqueName: \"kubernetes.io/projected/5ace24c1-074f-4c5b-8148-5bb46a790a3e-kube-api-access-fprvl\") pod \"auto-csr-approver-29556466-zhrhb\" (UID: \"5ace24c1-074f-4c5b-8148-5bb46a790a3e\") " pod="openshift-infra/auto-csr-approver-29556466-zhrhb" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.393013 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fprvl\" (UniqueName: \"kubernetes.io/projected/5ace24c1-074f-4c5b-8148-5bb46a790a3e-kube-api-access-fprvl\") pod \"auto-csr-approver-29556466-zhrhb\" (UID: \"5ace24c1-074f-4c5b-8148-5bb46a790a3e\") " pod="openshift-infra/auto-csr-approver-29556466-zhrhb" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.473427 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556466-zhrhb" Mar 13 07:46:00 crc kubenswrapper[4876]: I0313 07:46:00.941623 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556466-zhrhb"] Mar 13 07:46:01 crc kubenswrapper[4876]: I0313 07:46:01.507201 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556466-zhrhb" event={"ID":"5ace24c1-074f-4c5b-8148-5bb46a790a3e","Type":"ContainerStarted","Data":"00f923efa0369b9de0797ea197d914ef6ae078517179f1961af5f283c923e6eb"} Mar 13 07:46:02 crc kubenswrapper[4876]: I0313 07:46:02.519147 4876 generic.go:334] "Generic (PLEG): container finished" podID="5ace24c1-074f-4c5b-8148-5bb46a790a3e" containerID="670a49da83eeaa094431c2a1bd7be454953fd61b819d601b25a86ec9ed070bfb" exitCode=0 Mar 13 07:46:02 crc kubenswrapper[4876]: I0313 07:46:02.519231 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556466-zhrhb" event={"ID":"5ace24c1-074f-4c5b-8148-5bb46a790a3e","Type":"ContainerDied","Data":"670a49da83eeaa094431c2a1bd7be454953fd61b819d601b25a86ec9ed070bfb"} Mar 13 07:46:03 crc kubenswrapper[4876]: I0313 07:46:03.816355 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556466-zhrhb" Mar 13 07:46:03 crc kubenswrapper[4876]: I0313 07:46:03.924920 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fprvl\" (UniqueName: \"kubernetes.io/projected/5ace24c1-074f-4c5b-8148-5bb46a790a3e-kube-api-access-fprvl\") pod \"5ace24c1-074f-4c5b-8148-5bb46a790a3e\" (UID: \"5ace24c1-074f-4c5b-8148-5bb46a790a3e\") " Mar 13 07:46:03 crc kubenswrapper[4876]: I0313 07:46:03.934684 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ace24c1-074f-4c5b-8148-5bb46a790a3e-kube-api-access-fprvl" (OuterVolumeSpecName: "kube-api-access-fprvl") pod "5ace24c1-074f-4c5b-8148-5bb46a790a3e" (UID: "5ace24c1-074f-4c5b-8148-5bb46a790a3e"). InnerVolumeSpecName "kube-api-access-fprvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:46:04 crc kubenswrapper[4876]: I0313 07:46:04.027919 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fprvl\" (UniqueName: \"kubernetes.io/projected/5ace24c1-074f-4c5b-8148-5bb46a790a3e-kube-api-access-fprvl\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:04 crc kubenswrapper[4876]: I0313 07:46:04.537889 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556466-zhrhb" event={"ID":"5ace24c1-074f-4c5b-8148-5bb46a790a3e","Type":"ContainerDied","Data":"00f923efa0369b9de0797ea197d914ef6ae078517179f1961af5f283c923e6eb"} Mar 13 07:46:04 crc kubenswrapper[4876]: I0313 07:46:04.538449 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00f923efa0369b9de0797ea197d914ef6ae078517179f1961af5f283c923e6eb" Mar 13 07:46:04 crc kubenswrapper[4876]: I0313 07:46:04.537964 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556466-zhrhb" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.558395 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kgzdt"] Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.559517 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kgzdt" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="registry-server" containerID="cri-o://6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7" gracePeriod=30 Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.568519 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fmcl5"] Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.569027 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fmcl5" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerName="registry-server" containerID="cri-o://feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38" gracePeriod=30 Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.590735 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-75kbq"] Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.591060 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" podUID="f089ab79-2e29-4074-a8dc-21b0af79d3f7" containerName="marketplace-operator" containerID="cri-o://647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2" gracePeriod=30 Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.603393 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q97b"] Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.603852 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2q97b" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="registry-server" containerID="cri-o://24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346" gracePeriod=30 Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.610907 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzmpk"] Mar 13 07:46:09 crc kubenswrapper[4876]: E0313 07:46:09.611245 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ace24c1-074f-4c5b-8148-5bb46a790a3e" containerName="oc" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.611269 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ace24c1-074f-4c5b-8148-5bb46a790a3e" containerName="oc" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.611413 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ace24c1-074f-4c5b-8148-5bb46a790a3e" containerName="oc" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.611999 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.620101 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kdxtz"] Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.620479 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kdxtz" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerName="registry-server" containerID="cri-o://157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473" gracePeriod=30 Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.629515 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzmpk"] Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.710922 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0adab6c-7942-4756-829f-3b7a6569f8cb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.712197 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a0adab6c-7942-4756-829f-3b7a6569f8cb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.712300 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpf2t\" (UniqueName: \"kubernetes.io/projected/a0adab6c-7942-4756-829f-3b7a6569f8cb-kube-api-access-mpf2t\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.813379 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0adab6c-7942-4756-829f-3b7a6569f8cb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.813452 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a0adab6c-7942-4756-829f-3b7a6569f8cb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.813503 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpf2t\" (UniqueName: \"kubernetes.io/projected/a0adab6c-7942-4756-829f-3b7a6569f8cb-kube-api-access-mpf2t\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.814973 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0adab6c-7942-4756-829f-3b7a6569f8cb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.824395 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a0adab6c-7942-4756-829f-3b7a6569f8cb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:09 crc kubenswrapper[4876]: I0313 07:46:09.834056 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpf2t\" (UniqueName: \"kubernetes.io/projected/a0adab6c-7942-4756-829f-3b7a6569f8cb-kube-api-access-mpf2t\") pod \"marketplace-operator-79b997595-gzmpk\" (UID: \"a0adab6c-7942-4756-829f-3b7a6569f8cb\") " pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.067305 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.079019 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.175482 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.191410 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.222616 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-catalog-content\") pod \"af6d5722-1354-4021-9ba4-6ad9a1286c60\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.223759 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl8wt\" (UniqueName: \"kubernetes.io/projected/af6d5722-1354-4021-9ba4-6ad9a1286c60-kube-api-access-zl8wt\") pod \"af6d5722-1354-4021-9ba4-6ad9a1286c60\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.226168 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-utilities\") pod \"af6d5722-1354-4021-9ba4-6ad9a1286c60\" (UID: \"af6d5722-1354-4021-9ba4-6ad9a1286c60\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.229517 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-utilities" (OuterVolumeSpecName: "utilities") pod "af6d5722-1354-4021-9ba4-6ad9a1286c60" (UID: "af6d5722-1354-4021-9ba4-6ad9a1286c60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.234386 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af6d5722-1354-4021-9ba4-6ad9a1286c60-kube-api-access-zl8wt" (OuterVolumeSpecName: "kube-api-access-zl8wt") pod "af6d5722-1354-4021-9ba4-6ad9a1286c60" (UID: "af6d5722-1354-4021-9ba4-6ad9a1286c60"). InnerVolumeSpecName "kube-api-access-zl8wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.237028 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.238904 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.320916 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af6d5722-1354-4021-9ba4-6ad9a1286c60" (UID: "af6d5722-1354-4021-9ba4-6ad9a1286c60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328129 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-utilities\") pod \"63f3a5db-0db1-4486-a982-39c60bda9dc4\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328187 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h95dx\" (UniqueName: \"kubernetes.io/projected/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-kube-api-access-h95dx\") pod \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328288 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-utilities\") pod \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328317 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-catalog-content\") pod \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\" (UID: \"f7fbdc82-2033-4493-8f3a-ee17b88eb94e\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328337 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8qb\" (UniqueName: \"kubernetes.io/projected/63f3a5db-0db1-4486-a982-39c60bda9dc4-kube-api-access-ws8qb\") pod \"63f3a5db-0db1-4486-a982-39c60bda9dc4\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328402 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-catalog-content\") pod \"63f3a5db-0db1-4486-a982-39c60bda9dc4\" (UID: \"63f3a5db-0db1-4486-a982-39c60bda9dc4\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328856 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl8wt\" (UniqueName: \"kubernetes.io/projected/af6d5722-1354-4021-9ba4-6ad9a1286c60-kube-api-access-zl8wt\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328879 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.328889 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6d5722-1354-4021-9ba4-6ad9a1286c60-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.329775 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-utilities" (OuterVolumeSpecName: "utilities") pod "63f3a5db-0db1-4486-a982-39c60bda9dc4" (UID: "63f3a5db-0db1-4486-a982-39c60bda9dc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.330064 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-utilities" (OuterVolumeSpecName: "utilities") pod "f7fbdc82-2033-4493-8f3a-ee17b88eb94e" (UID: "f7fbdc82-2033-4493-8f3a-ee17b88eb94e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.333471 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63f3a5db-0db1-4486-a982-39c60bda9dc4-kube-api-access-ws8qb" (OuterVolumeSpecName: "kube-api-access-ws8qb") pod "63f3a5db-0db1-4486-a982-39c60bda9dc4" (UID: "63f3a5db-0db1-4486-a982-39c60bda9dc4"). InnerVolumeSpecName "kube-api-access-ws8qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.333596 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-kube-api-access-h95dx" (OuterVolumeSpecName: "kube-api-access-h95dx") pod "f7fbdc82-2033-4493-8f3a-ee17b88eb94e" (UID: "f7fbdc82-2033-4493-8f3a-ee17b88eb94e"). InnerVolumeSpecName "kube-api-access-h95dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.356485 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63f3a5db-0db1-4486-a982-39c60bda9dc4" (UID: "63f3a5db-0db1-4486-a982-39c60bda9dc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.429789 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-utilities\") pod \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.429941 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qppc7\" (UniqueName: \"kubernetes.io/projected/f089ab79-2e29-4074-a8dc-21b0af79d3f7-kube-api-access-qppc7\") pod \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.429990 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-trusted-ca\") pod \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430091 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-catalog-content\") pod \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430198 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-operator-metrics\") pod \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\" (UID: \"f089ab79-2e29-4074-a8dc-21b0af79d3f7\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430258 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ld7c\" (UniqueName: \"kubernetes.io/projected/8c486d8f-87cf-42c0-9b8c-8f715755edc8-kube-api-access-9ld7c\") pod \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\" (UID: \"8c486d8f-87cf-42c0-9b8c-8f715755edc8\") " Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430495 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430508 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63f3a5db-0db1-4486-a982-39c60bda9dc4-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430519 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h95dx\" (UniqueName: \"kubernetes.io/projected/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-kube-api-access-h95dx\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430531 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430541 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws8qb\" (UniqueName: \"kubernetes.io/projected/63f3a5db-0db1-4486-a982-39c60bda9dc4-kube-api-access-ws8qb\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430854 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-utilities" (OuterVolumeSpecName: "utilities") pod "8c486d8f-87cf-42c0-9b8c-8f715755edc8" (UID: "8c486d8f-87cf-42c0-9b8c-8f715755edc8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.430871 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f089ab79-2e29-4074-a8dc-21b0af79d3f7" (UID: "f089ab79-2e29-4074-a8dc-21b0af79d3f7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.434324 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f089ab79-2e29-4074-a8dc-21b0af79d3f7-kube-api-access-qppc7" (OuterVolumeSpecName: "kube-api-access-qppc7") pod "f089ab79-2e29-4074-a8dc-21b0af79d3f7" (UID: "f089ab79-2e29-4074-a8dc-21b0af79d3f7"). InnerVolumeSpecName "kube-api-access-qppc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.434389 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c486d8f-87cf-42c0-9b8c-8f715755edc8-kube-api-access-9ld7c" (OuterVolumeSpecName: "kube-api-access-9ld7c") pod "8c486d8f-87cf-42c0-9b8c-8f715755edc8" (UID: "8c486d8f-87cf-42c0-9b8c-8f715755edc8"). InnerVolumeSpecName "kube-api-access-9ld7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.435282 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f089ab79-2e29-4074-a8dc-21b0af79d3f7" (UID: "f089ab79-2e29-4074-a8dc-21b0af79d3f7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.461173 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7fbdc82-2033-4493-8f3a-ee17b88eb94e" (UID: "f7fbdc82-2033-4493-8f3a-ee17b88eb94e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.494857 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c486d8f-87cf-42c0-9b8c-8f715755edc8" (UID: "8c486d8f-87cf-42c0-9b8c-8f715755edc8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.532461 4876 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.532512 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ld7c\" (UniqueName: \"kubernetes.io/projected/8c486d8f-87cf-42c0-9b8c-8f715755edc8-kube-api-access-9ld7c\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.532523 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.532534 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7fbdc82-2033-4493-8f3a-ee17b88eb94e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.532548 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qppc7\" (UniqueName: \"kubernetes.io/projected/f089ab79-2e29-4074-a8dc-21b0af79d3f7-kube-api-access-qppc7\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.532557 4876 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f089ab79-2e29-4074-a8dc-21b0af79d3f7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.532566 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c486d8f-87cf-42c0-9b8c-8f715755edc8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.589338 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gzmpk"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.594392 4876 generic.go:334] "Generic (PLEG): container finished" podID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerID="6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7" exitCode=0 Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.594448 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kgzdt" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.594496 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgzdt" event={"ID":"af6d5722-1354-4021-9ba4-6ad9a1286c60","Type":"ContainerDied","Data":"6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.594556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kgzdt" event={"ID":"af6d5722-1354-4021-9ba4-6ad9a1286c60","Type":"ContainerDied","Data":"486d7fdf7bb9b66bb0104a13afa68c6e8110ff958605093e669b58880a6d6c00"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.594657 4876 scope.go:117] "RemoveContainer" containerID="6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.598148 4876 generic.go:334] "Generic (PLEG): container finished" podID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerID="157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473" exitCode=0 Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.598320 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdxtz" event={"ID":"f7fbdc82-2033-4493-8f3a-ee17b88eb94e","Type":"ContainerDied","Data":"157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.598362 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdxtz" event={"ID":"f7fbdc82-2033-4493-8f3a-ee17b88eb94e","Type":"ContainerDied","Data":"e272a0aa889f2f6688e65c378f1e24df44a9de205b08e05e5972dd74aac96bbc"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.598467 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdxtz" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.601550 4876 generic.go:334] "Generic (PLEG): container finished" podID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerID="24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346" exitCode=0 Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.601671 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q97b" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.601882 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q97b" event={"ID":"63f3a5db-0db1-4486-a982-39c60bda9dc4","Type":"ContainerDied","Data":"24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.601920 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q97b" event={"ID":"63f3a5db-0db1-4486-a982-39c60bda9dc4","Type":"ContainerDied","Data":"521b126297d22b9a1389665d013f1e667e8f021cdccbbbc29b37c570afdab435"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.605295 4876 generic.go:334] "Generic (PLEG): container finished" podID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerID="feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38" exitCode=0 Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.605369 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcl5" event={"ID":"8c486d8f-87cf-42c0-9b8c-8f715755edc8","Type":"ContainerDied","Data":"feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.605406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcl5" event={"ID":"8c486d8f-87cf-42c0-9b8c-8f715755edc8","Type":"ContainerDied","Data":"6bfe74f805be6ad373fc748949952661b08729715e854d215c05420ca1ee568f"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.605540 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmcl5" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.608174 4876 generic.go:334] "Generic (PLEG): container finished" podID="f089ab79-2e29-4074-a8dc-21b0af79d3f7" containerID="647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2" exitCode=0 Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.608197 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.608212 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" event={"ID":"f089ab79-2e29-4074-a8dc-21b0af79d3f7","Type":"ContainerDied","Data":"647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.608463 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-75kbq" event={"ID":"f089ab79-2e29-4074-a8dc-21b0af79d3f7","Type":"ContainerDied","Data":"052a6b45f498fd6d280d3e717ed76cb57f975004a439506fd4c7b4ab3f00d161"} Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.628327 4876 scope.go:117] "RemoveContainer" containerID="e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.637312 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kgzdt"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.643722 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kgzdt"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.656551 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q97b"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.659737 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q97b"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.668778 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kdxtz"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.682572 4876 scope.go:117] "RemoveContainer" containerID="ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.686813 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kdxtz"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.694733 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-75kbq"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.709707 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-75kbq"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.711420 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fmcl5"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.714630 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fmcl5"] Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.771026 4876 scope.go:117] "RemoveContainer" containerID="6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.772374 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7\": container with ID starting with 6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7 not found: ID does not exist" containerID="6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.772434 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7"} err="failed to get container status \"6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7\": rpc error: code = NotFound desc = could not find container \"6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7\": container with ID starting with 6c4766f81dc33a90cdb0827d259b8b70dd0b1d2bf7ee92694a8552dd1b0908f7 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.772480 4876 scope.go:117] "RemoveContainer" containerID="e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.773279 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a\": container with ID starting with e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a not found: ID does not exist" containerID="e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.773301 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a"} err="failed to get container status \"e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a\": rpc error: code = NotFound desc = could not find container \"e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a\": container with ID starting with e298b87acf99ad606f2d0e9b51f7285a2d689a27a0651d5e96e56b7db8ebe51a not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.773319 4876 scope.go:117] "RemoveContainer" containerID="ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.773788 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc\": container with ID starting with ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc not found: ID does not exist" containerID="ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.773811 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc"} err="failed to get container status \"ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc\": rpc error: code = NotFound desc = could not find container \"ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc\": container with ID starting with ef6aa96910aaae42099c3f05a9f6380cf7ed2e8e1587d502e3ce5e60e1e8d7fc not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.773823 4876 scope.go:117] "RemoveContainer" containerID="157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.791581 4876 scope.go:117] "RemoveContainer" containerID="385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.827484 4876 scope.go:117] "RemoveContainer" containerID="aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.852636 4876 scope.go:117] "RemoveContainer" containerID="157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.854828 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473\": container with ID starting with 157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473 not found: ID does not exist" containerID="157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.854895 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473"} err="failed to get container status \"157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473\": rpc error: code = NotFound desc = could not find container \"157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473\": container with ID starting with 157ee5613f6ea5369e3c9cd708276990d0e7238b91d4dd1cbf5ff2a27351e473 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.854935 4876 scope.go:117] "RemoveContainer" containerID="385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.855527 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463\": container with ID starting with 385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463 not found: ID does not exist" containerID="385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.855582 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463"} err="failed to get container status \"385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463\": rpc error: code = NotFound desc = could not find container \"385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463\": container with ID starting with 385bbdc04b5fd1954b10deb587d73c6755f2f6504795ffdb71a4707dff079463 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.855611 4876 scope.go:117] "RemoveContainer" containerID="aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.856087 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62\": container with ID starting with aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62 not found: ID does not exist" containerID="aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.856157 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62"} err="failed to get container status \"aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62\": rpc error: code = NotFound desc = could not find container \"aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62\": container with ID starting with aebfd58aa5e360badfb77e1c26547d8c646469dc4ab996ccf3a3649fba42ee62 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.856195 4876 scope.go:117] "RemoveContainer" containerID="24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.881936 4876 scope.go:117] "RemoveContainer" containerID="df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.895757 4876 scope.go:117] "RemoveContainer" containerID="c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.922615 4876 scope.go:117] "RemoveContainer" containerID="24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.923366 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346\": container with ID starting with 24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346 not found: ID does not exist" containerID="24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.923443 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346"} err="failed to get container status \"24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346\": rpc error: code = NotFound desc = could not find container \"24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346\": container with ID starting with 24dfa22c4ffc8645a5361eb37eb6aa7ee5076137b2f6f8b39e2bb7a43a6e8346 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.923488 4876 scope.go:117] "RemoveContainer" containerID="df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.923958 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92\": container with ID starting with df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92 not found: ID does not exist" containerID="df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.924012 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92"} err="failed to get container status \"df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92\": rpc error: code = NotFound desc = could not find container \"df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92\": container with ID starting with df572ad089722d49f849175eedda352a41b0b292a0e3eecb05c57ea1711dee92 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.924053 4876 scope.go:117] "RemoveContainer" containerID="c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.924462 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2\": container with ID starting with c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2 not found: ID does not exist" containerID="c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.924498 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2"} err="failed to get container status \"c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2\": rpc error: code = NotFound desc = could not find container \"c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2\": container with ID starting with c5a36925469b6f4b1714fed8c71b031858f9ca5d5823248a4654585f9a47a9b2 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.924518 4876 scope.go:117] "RemoveContainer" containerID="feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.940800 4876 scope.go:117] "RemoveContainer" containerID="5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.965430 4876 scope.go:117] "RemoveContainer" containerID="44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.982430 4876 scope.go:117] "RemoveContainer" containerID="feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.983080 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38\": container with ID starting with feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38 not found: ID does not exist" containerID="feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.983146 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38"} err="failed to get container status \"feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38\": rpc error: code = NotFound desc = could not find container \"feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38\": container with ID starting with feebb8e45ceb1844fefddf3c5dbe6bdb49a113afa43149dcbd1c463ab06e6d38 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.983186 4876 scope.go:117] "RemoveContainer" containerID="5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.983607 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033\": container with ID starting with 5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033 not found: ID does not exist" containerID="5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.983664 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033"} err="failed to get container status \"5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033\": rpc error: code = NotFound desc = could not find container \"5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033\": container with ID starting with 5b2b1ec573cc0e117e347e1f4abd302d8f508afbda7aeec7654769b103ed6033 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.983704 4876 scope.go:117] "RemoveContainer" containerID="44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24" Mar 13 07:46:10 crc kubenswrapper[4876]: E0313 07:46:10.984043 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24\": container with ID starting with 44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24 not found: ID does not exist" containerID="44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.984074 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24"} err="failed to get container status \"44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24\": rpc error: code = NotFound desc = could not find container \"44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24\": container with ID starting with 44c944cddc7b970ebfdff5a8fdd71fdeb8a21632964ef77f1b900a18a7c54f24 not found: ID does not exist" Mar 13 07:46:10 crc kubenswrapper[4876]: I0313 07:46:10.984090 4876 scope.go:117] "RemoveContainer" containerID="647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.000640 4876 scope.go:117] "RemoveContainer" containerID="647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.001345 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2\": container with ID starting with 647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2 not found: ID does not exist" containerID="647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.001411 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2"} err="failed to get container status \"647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2\": rpc error: code = NotFound desc = could not find container \"647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2\": container with ID starting with 647926afd6f97bf86bec880bcc7565a98d2604e845870bf147f592e797d4f4e2 not found: ID does not exist" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.042150 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" path="/var/lib/kubelet/pods/63f3a5db-0db1-4486-a982-39c60bda9dc4/volumes" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.043485 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" path="/var/lib/kubelet/pods/8c486d8f-87cf-42c0-9b8c-8f715755edc8/volumes" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.044199 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" path="/var/lib/kubelet/pods/af6d5722-1354-4021-9ba4-6ad9a1286c60/volumes" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.045422 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f089ab79-2e29-4074-a8dc-21b0af79d3f7" path="/var/lib/kubelet/pods/f089ab79-2e29-4074-a8dc-21b0af79d3f7/volumes" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.045970 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" path="/var/lib/kubelet/pods/f7fbdc82-2033-4493-8f3a-ee17b88eb94e/volumes" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.344736 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gkkq5"] Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345006 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerName="extract-utilities" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345020 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerName="extract-utilities" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345030 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="extract-content" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345036 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="extract-content" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345044 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerName="extract-content" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345050 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerName="extract-content" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345062 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="extract-utilities" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345067 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="extract-utilities" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345075 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerName="extract-utilities" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345082 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerName="extract-utilities" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345092 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="extract-content" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345098 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="extract-content" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345106 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345111 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345119 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f089ab79-2e29-4074-a8dc-21b0af79d3f7" containerName="marketplace-operator" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345124 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f089ab79-2e29-4074-a8dc-21b0af79d3f7" containerName="marketplace-operator" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345132 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerName="extract-content" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345138 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerName="extract-content" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345147 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345153 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345160 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345168 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345175 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345181 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: E0313 07:46:11.345194 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="extract-utilities" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345200 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="extract-utilities" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345322 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="af6d5722-1354-4021-9ba4-6ad9a1286c60" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345342 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7fbdc82-2033-4493-8f3a-ee17b88eb94e" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345349 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c486d8f-87cf-42c0-9b8c-8f715755edc8" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345356 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f089ab79-2e29-4074-a8dc-21b0af79d3f7" containerName="marketplace-operator" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.345364 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="63f3a5db-0db1-4486-a982-39c60bda9dc4" containerName="registry-server" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.346147 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.350119 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.358899 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkkq5"] Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.444395 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-utilities\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.444455 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkmrs\" (UniqueName: \"kubernetes.io/projected/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-kube-api-access-wkmrs\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.444503 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-catalog-content\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.545969 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-catalog-content\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.546490 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-utilities\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.546661 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkmrs\" (UniqueName: \"kubernetes.io/projected/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-kube-api-access-wkmrs\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.547035 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-catalog-content\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.547112 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-utilities\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.570668 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkmrs\" (UniqueName: \"kubernetes.io/projected/cccd9f80-f78e-4e2b-a101-d85fa5b8a942-kube-api-access-wkmrs\") pod \"redhat-marketplace-gkkq5\" (UID: \"cccd9f80-f78e-4e2b-a101-d85fa5b8a942\") " pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.621152 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" event={"ID":"a0adab6c-7942-4756-829f-3b7a6569f8cb","Type":"ContainerStarted","Data":"31918e2a96caa540e4a0baa42d0ee460507d29f2e0b4f85e5ce379863593a10c"} Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.621204 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" event={"ID":"a0adab6c-7942-4756-829f-3b7a6569f8cb","Type":"ContainerStarted","Data":"a403cf31a4131a198aa1ff8c561f682f126374684384794d402d47e52ec538fa"} Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.621472 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.624850 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.640446 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gzmpk" podStartSLOduration=2.640420096 podStartE2EDuration="2.640420096s" podCreationTimestamp="2026-03-13 07:46:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:46:11.635688046 +0000 UTC m=+431.306467028" watchObservedRunningTime="2026-03-13 07:46:11.640420096 +0000 UTC m=+431.311199078" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.671737 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:11 crc kubenswrapper[4876]: I0313 07:46:11.910656 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkkq5"] Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.352371 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zvfb8"] Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.354084 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.355278 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zvfb8"] Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.357473 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.462876 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-catalog-content\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.463018 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hch2f\" (UniqueName: \"kubernetes.io/projected/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-kube-api-access-hch2f\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.463092 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-utilities\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.565059 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-utilities\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.565200 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-catalog-content\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.565314 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hch2f\" (UniqueName: \"kubernetes.io/projected/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-kube-api-access-hch2f\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.565945 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-utilities\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.566091 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-catalog-content\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.594309 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hch2f\" (UniqueName: \"kubernetes.io/projected/729a14e7-a89b-4e7c-a8b5-a4002519d6a3-kube-api-access-hch2f\") pod \"redhat-operators-zvfb8\" (UID: \"729a14e7-a89b-4e7c-a8b5-a4002519d6a3\") " pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.631014 4876 generic.go:334] "Generic (PLEG): container finished" podID="cccd9f80-f78e-4e2b-a101-d85fa5b8a942" containerID="f26a6c33e86db967f013cd8b938ab1e563a3be4b709336842f6e9f0364555966" exitCode=0 Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.631150 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkkq5" event={"ID":"cccd9f80-f78e-4e2b-a101-d85fa5b8a942","Type":"ContainerDied","Data":"f26a6c33e86db967f013cd8b938ab1e563a3be4b709336842f6e9f0364555966"} Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.631299 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkkq5" event={"ID":"cccd9f80-f78e-4e2b-a101-d85fa5b8a942","Type":"ContainerStarted","Data":"a68c281c8ad6e739b84014d2aca693c1564cc7ccee40418c13dcde259d7552fa"} Mar 13 07:46:12 crc kubenswrapper[4876]: I0313 07:46:12.675753 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:13 crc kubenswrapper[4876]: W0313 07:46:13.155180 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod729a14e7_a89b_4e7c_a8b5_a4002519d6a3.slice/crio-211361bd989a2455c3f18ec8b1526b34ab700aa4156b7a855aa9b7613d3665e3 WatchSource:0}: Error finding container 211361bd989a2455c3f18ec8b1526b34ab700aa4156b7a855aa9b7613d3665e3: Status 404 returned error can't find the container with id 211361bd989a2455c3f18ec8b1526b34ab700aa4156b7a855aa9b7613d3665e3 Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.155715 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zvfb8"] Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.643077 4876 generic.go:334] "Generic (PLEG): container finished" podID="729a14e7-a89b-4e7c-a8b5-a4002519d6a3" containerID="89d84f71d2039c0ea955c7777b847983908d891743a2f5b245edb1bfeb9d5cf9" exitCode=0 Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.643180 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvfb8" event={"ID":"729a14e7-a89b-4e7c-a8b5-a4002519d6a3","Type":"ContainerDied","Data":"89d84f71d2039c0ea955c7777b847983908d891743a2f5b245edb1bfeb9d5cf9"} Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.643218 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvfb8" event={"ID":"729a14e7-a89b-4e7c-a8b5-a4002519d6a3","Type":"ContainerStarted","Data":"211361bd989a2455c3f18ec8b1526b34ab700aa4156b7a855aa9b7613d3665e3"} Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.648289 4876 generic.go:334] "Generic (PLEG): container finished" podID="cccd9f80-f78e-4e2b-a101-d85fa5b8a942" containerID="e8d1dd81b7d9b36769afd3a05e1cb2ad5965671505cb53cf2d345d387b00654c" exitCode=0 Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.648693 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkkq5" event={"ID":"cccd9f80-f78e-4e2b-a101-d85fa5b8a942","Type":"ContainerDied","Data":"e8d1dd81b7d9b36769afd3a05e1cb2ad5965671505cb53cf2d345d387b00654c"} Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.750108 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8rcs9"] Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.751601 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.757130 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.758990 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rcs9"] Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.784911 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjns7\" (UniqueName: \"kubernetes.io/projected/0e320bf9-7cd4-423f-af2a-9e00e602bcff-kube-api-access-mjns7\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.784980 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e320bf9-7cd4-423f-af2a-9e00e602bcff-utilities\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.785013 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e320bf9-7cd4-423f-af2a-9e00e602bcff-catalog-content\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.886676 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjns7\" (UniqueName: \"kubernetes.io/projected/0e320bf9-7cd4-423f-af2a-9e00e602bcff-kube-api-access-mjns7\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.886799 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e320bf9-7cd4-423f-af2a-9e00e602bcff-utilities\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.886843 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e320bf9-7cd4-423f-af2a-9e00e602bcff-catalog-content\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.887657 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e320bf9-7cd4-423f-af2a-9e00e602bcff-utilities\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.887724 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e320bf9-7cd4-423f-af2a-9e00e602bcff-catalog-content\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:13 crc kubenswrapper[4876]: I0313 07:46:13.912535 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjns7\" (UniqueName: \"kubernetes.io/projected/0e320bf9-7cd4-423f-af2a-9e00e602bcff-kube-api-access-mjns7\") pod \"certified-operators-8rcs9\" (UID: \"0e320bf9-7cd4-423f-af2a-9e00e602bcff\") " pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.072789 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.550759 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rcs9"] Mar 13 07:46:14 crc kubenswrapper[4876]: W0313 07:46:14.551085 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e320bf9_7cd4_423f_af2a_9e00e602bcff.slice/crio-fef78d8530b468f1013d7dccf77d603ae761fcd890c2b2d9ec697f9820fe65b6 WatchSource:0}: Error finding container fef78d8530b468f1013d7dccf77d603ae761fcd890c2b2d9ec697f9820fe65b6: Status 404 returned error can't find the container with id fef78d8530b468f1013d7dccf77d603ae761fcd890c2b2d9ec697f9820fe65b6 Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.663129 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkkq5" event={"ID":"cccd9f80-f78e-4e2b-a101-d85fa5b8a942","Type":"ContainerStarted","Data":"e4a83820217768f7962ed39119ee631db17761ef4cc9209194f337599429def4"} Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.670030 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rcs9" event={"ID":"0e320bf9-7cd4-423f-af2a-9e00e602bcff","Type":"ContainerStarted","Data":"fef78d8530b468f1013d7dccf77d603ae761fcd890c2b2d9ec697f9820fe65b6"} Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.672086 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvfb8" event={"ID":"729a14e7-a89b-4e7c-a8b5-a4002519d6a3","Type":"ContainerStarted","Data":"b281739d2a958dd791e538fe7628b1a0173e9c970f77f2e6239b4760e7b6c988"} Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.691344 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gkkq5" podStartSLOduration=2.272911695 podStartE2EDuration="3.691316001s" podCreationTimestamp="2026-03-13 07:46:11 +0000 UTC" firstStartedPulling="2026-03-13 07:46:12.632808121 +0000 UTC m=+432.303587103" lastFinishedPulling="2026-03-13 07:46:14.051212427 +0000 UTC m=+433.721991409" observedRunningTime="2026-03-13 07:46:14.68994217 +0000 UTC m=+434.360721162" watchObservedRunningTime="2026-03-13 07:46:14.691316001 +0000 UTC m=+434.362094983" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.740553 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pxfhc"] Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.744277 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.746650 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.763832 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxfhc"] Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.798677 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-utilities\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.798755 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxv4m\" (UniqueName: \"kubernetes.io/projected/add82187-d01b-4902-8f12-ced45327d1f9-kube-api-access-pxv4m\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.798894 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-catalog-content\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.899890 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-catalog-content\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.900380 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-utilities\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.900511 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxv4m\" (UniqueName: \"kubernetes.io/projected/add82187-d01b-4902-8f12-ced45327d1f9-kube-api-access-pxv4m\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.900461 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-catalog-content\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.900651 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-utilities\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:14 crc kubenswrapper[4876]: I0313 07:46:14.926689 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxv4m\" (UniqueName: \"kubernetes.io/projected/add82187-d01b-4902-8f12-ced45327d1f9-kube-api-access-pxv4m\") pod \"community-operators-pxfhc\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:15 crc kubenswrapper[4876]: I0313 07:46:15.058069 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:15 crc kubenswrapper[4876]: I0313 07:46:15.542180 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxfhc"] Mar 13 07:46:15 crc kubenswrapper[4876]: I0313 07:46:15.679910 4876 generic.go:334] "Generic (PLEG): container finished" podID="729a14e7-a89b-4e7c-a8b5-a4002519d6a3" containerID="b281739d2a958dd791e538fe7628b1a0173e9c970f77f2e6239b4760e7b6c988" exitCode=0 Mar 13 07:46:15 crc kubenswrapper[4876]: I0313 07:46:15.680004 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvfb8" event={"ID":"729a14e7-a89b-4e7c-a8b5-a4002519d6a3","Type":"ContainerDied","Data":"b281739d2a958dd791e538fe7628b1a0173e9c970f77f2e6239b4760e7b6c988"} Mar 13 07:46:15 crc kubenswrapper[4876]: I0313 07:46:15.681365 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxfhc" event={"ID":"add82187-d01b-4902-8f12-ced45327d1f9","Type":"ContainerStarted","Data":"decc1a185f5ebc48cf055b8bda5f932c702c82617187fadb1cac20e501d8d29e"} Mar 13 07:46:15 crc kubenswrapper[4876]: I0313 07:46:15.683455 4876 generic.go:334] "Generic (PLEG): container finished" podID="0e320bf9-7cd4-423f-af2a-9e00e602bcff" containerID="ecdb6d46f0a93cebc2dd980b445dc75d0ff126fcb0cc0c1f226c839d7c8c57b4" exitCode=0 Mar 13 07:46:15 crc kubenswrapper[4876]: I0313 07:46:15.685154 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rcs9" event={"ID":"0e320bf9-7cd4-423f-af2a-9e00e602bcff","Type":"ContainerDied","Data":"ecdb6d46f0a93cebc2dd980b445dc75d0ff126fcb0cc0c1f226c839d7c8c57b4"} Mar 13 07:46:16 crc kubenswrapper[4876]: I0313 07:46:16.697801 4876 generic.go:334] "Generic (PLEG): container finished" podID="0e320bf9-7cd4-423f-af2a-9e00e602bcff" containerID="50a6ea103ca993dc2afb88d99d2ad36c2b85e84ff0bb71d6e8f20ace11ae2f4a" exitCode=0 Mar 13 07:46:16 crc kubenswrapper[4876]: I0313 07:46:16.698359 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rcs9" event={"ID":"0e320bf9-7cd4-423f-af2a-9e00e602bcff","Type":"ContainerDied","Data":"50a6ea103ca993dc2afb88d99d2ad36c2b85e84ff0bb71d6e8f20ace11ae2f4a"} Mar 13 07:46:16 crc kubenswrapper[4876]: I0313 07:46:16.702183 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvfb8" event={"ID":"729a14e7-a89b-4e7c-a8b5-a4002519d6a3","Type":"ContainerStarted","Data":"7e28b8d37b5bb511468aae9a7d4c7abe899a090f68d5172f8943e7512cb05885"} Mar 13 07:46:16 crc kubenswrapper[4876]: I0313 07:46:16.705625 4876 generic.go:334] "Generic (PLEG): container finished" podID="add82187-d01b-4902-8f12-ced45327d1f9" containerID="2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511" exitCode=0 Mar 13 07:46:16 crc kubenswrapper[4876]: I0313 07:46:16.705676 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxfhc" event={"ID":"add82187-d01b-4902-8f12-ced45327d1f9","Type":"ContainerDied","Data":"2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511"} Mar 13 07:46:16 crc kubenswrapper[4876]: I0313 07:46:16.776472 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zvfb8" podStartSLOduration=2.267987779 podStartE2EDuration="4.776435854s" podCreationTimestamp="2026-03-13 07:46:12 +0000 UTC" firstStartedPulling="2026-03-13 07:46:13.645277037 +0000 UTC m=+433.316056019" lastFinishedPulling="2026-03-13 07:46:16.153725112 +0000 UTC m=+435.824504094" observedRunningTime="2026-03-13 07:46:16.773605901 +0000 UTC m=+436.444384903" watchObservedRunningTime="2026-03-13 07:46:16.776435854 +0000 UTC m=+436.447214866" Mar 13 07:46:17 crc kubenswrapper[4876]: I0313 07:46:17.720984 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rcs9" event={"ID":"0e320bf9-7cd4-423f-af2a-9e00e602bcff","Type":"ContainerStarted","Data":"555163355e0497f7d33ed3ea5b128b074df972a2c6ba68f3b185533c89c93ce5"} Mar 13 07:46:17 crc kubenswrapper[4876]: I0313 07:46:17.725623 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxfhc" event={"ID":"add82187-d01b-4902-8f12-ced45327d1f9","Type":"ContainerStarted","Data":"c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d"} Mar 13 07:46:17 crc kubenswrapper[4876]: I0313 07:46:17.745084 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8rcs9" podStartSLOduration=3.315405922 podStartE2EDuration="4.745054959s" podCreationTimestamp="2026-03-13 07:46:13 +0000 UTC" firstStartedPulling="2026-03-13 07:46:15.685889059 +0000 UTC m=+435.356668041" lastFinishedPulling="2026-03-13 07:46:17.115538076 +0000 UTC m=+436.786317078" observedRunningTime="2026-03-13 07:46:17.743766531 +0000 UTC m=+437.414545513" watchObservedRunningTime="2026-03-13 07:46:17.745054959 +0000 UTC m=+437.415833971" Mar 13 07:46:18 crc kubenswrapper[4876]: I0313 07:46:18.734712 4876 generic.go:334] "Generic (PLEG): container finished" podID="add82187-d01b-4902-8f12-ced45327d1f9" containerID="c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d" exitCode=0 Mar 13 07:46:18 crc kubenswrapper[4876]: I0313 07:46:18.734740 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxfhc" event={"ID":"add82187-d01b-4902-8f12-ced45327d1f9","Type":"ContainerDied","Data":"c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d"} Mar 13 07:46:19 crc kubenswrapper[4876]: I0313 07:46:19.742791 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxfhc" event={"ID":"add82187-d01b-4902-8f12-ced45327d1f9","Type":"ContainerStarted","Data":"18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e"} Mar 13 07:46:19 crc kubenswrapper[4876]: I0313 07:46:19.771377 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pxfhc" podStartSLOduration=3.209097221 podStartE2EDuration="5.771350561s" podCreationTimestamp="2026-03-13 07:46:14 +0000 UTC" firstStartedPulling="2026-03-13 07:46:16.710673718 +0000 UTC m=+436.381452720" lastFinishedPulling="2026-03-13 07:46:19.272927078 +0000 UTC m=+438.943706060" observedRunningTime="2026-03-13 07:46:19.766722005 +0000 UTC m=+439.437501017" watchObservedRunningTime="2026-03-13 07:46:19.771350561 +0000 UTC m=+439.442129563" Mar 13 07:46:21 crc kubenswrapper[4876]: I0313 07:46:21.673494 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:21 crc kubenswrapper[4876]: I0313 07:46:21.673597 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:21 crc kubenswrapper[4876]: I0313 07:46:21.735445 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:21 crc kubenswrapper[4876]: I0313 07:46:21.807987 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gkkq5" Mar 13 07:46:22 crc kubenswrapper[4876]: I0313 07:46:22.677545 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:22 crc kubenswrapper[4876]: I0313 07:46:22.677641 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:23 crc kubenswrapper[4876]: I0313 07:46:23.742178 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zvfb8" podUID="729a14e7-a89b-4e7c-a8b5-a4002519d6a3" containerName="registry-server" probeResult="failure" output=< Mar 13 07:46:23 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 07:46:23 crc kubenswrapper[4876]: > Mar 13 07:46:24 crc kubenswrapper[4876]: I0313 07:46:24.073221 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:24 crc kubenswrapper[4876]: I0313 07:46:24.073327 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:24 crc kubenswrapper[4876]: I0313 07:46:24.132393 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:24 crc kubenswrapper[4876]: I0313 07:46:24.608045 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:46:24 crc kubenswrapper[4876]: I0313 07:46:24.608166 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:46:24 crc kubenswrapper[4876]: I0313 07:46:24.832467 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8rcs9" Mar 13 07:46:25 crc kubenswrapper[4876]: I0313 07:46:25.058528 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:25 crc kubenswrapper[4876]: I0313 07:46:25.058604 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:25 crc kubenswrapper[4876]: I0313 07:46:25.117608 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:25 crc kubenswrapper[4876]: I0313 07:46:25.833540 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pxfhc" Mar 13 07:46:32 crc kubenswrapper[4876]: I0313 07:46:32.749810 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:32 crc kubenswrapper[4876]: I0313 07:46:32.827562 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zvfb8" Mar 13 07:46:54 crc kubenswrapper[4876]: I0313 07:46:54.607683 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:46:54 crc kubenswrapper[4876]: I0313 07:46:54.608890 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:47:24 crc kubenswrapper[4876]: I0313 07:47:24.608380 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:47:24 crc kubenswrapper[4876]: I0313 07:47:24.609145 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:47:24 crc kubenswrapper[4876]: I0313 07:47:24.609285 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:47:24 crc kubenswrapper[4876]: I0313 07:47:24.610414 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f60a1005aa99b7a7a6f69d8ddd25eddf167d10b844bb62bc7097f7905502e52"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 07:47:24 crc kubenswrapper[4876]: I0313 07:47:24.610510 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://9f60a1005aa99b7a7a6f69d8ddd25eddf167d10b844bb62bc7097f7905502e52" gracePeriod=600 Mar 13 07:47:25 crc kubenswrapper[4876]: I0313 07:47:25.229542 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="9f60a1005aa99b7a7a6f69d8ddd25eddf167d10b844bb62bc7097f7905502e52" exitCode=0 Mar 13 07:47:25 crc kubenswrapper[4876]: I0313 07:47:25.229869 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"9f60a1005aa99b7a7a6f69d8ddd25eddf167d10b844bb62bc7097f7905502e52"} Mar 13 07:47:25 crc kubenswrapper[4876]: I0313 07:47:25.230136 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"ba651ad626e0def50adb9d0b266fb9ee781c393b76906dacbf80c73b1a02e496"} Mar 13 07:47:25 crc kubenswrapper[4876]: I0313 07:47:25.230169 4876 scope.go:117] "RemoveContainer" containerID="4476c64897b2339e4fbabc48ebb22eb4bb4e2c90d7f00809288cbeb47c9acc4c" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.147736 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556468-njzx8"] Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.149379 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556468-njzx8" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.154370 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.154903 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556468-njzx8"] Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.155285 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.159710 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.289275 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt8qj\" (UniqueName: \"kubernetes.io/projected/837c43f4-e275-4668-b5c2-34f7eead6093-kube-api-access-wt8qj\") pod \"auto-csr-approver-29556468-njzx8\" (UID: \"837c43f4-e275-4668-b5c2-34f7eead6093\") " pod="openshift-infra/auto-csr-approver-29556468-njzx8" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.390154 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt8qj\" (UniqueName: \"kubernetes.io/projected/837c43f4-e275-4668-b5c2-34f7eead6093-kube-api-access-wt8qj\") pod \"auto-csr-approver-29556468-njzx8\" (UID: \"837c43f4-e275-4668-b5c2-34f7eead6093\") " pod="openshift-infra/auto-csr-approver-29556468-njzx8" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.416804 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt8qj\" (UniqueName: \"kubernetes.io/projected/837c43f4-e275-4668-b5c2-34f7eead6093-kube-api-access-wt8qj\") pod \"auto-csr-approver-29556468-njzx8\" (UID: \"837c43f4-e275-4668-b5c2-34f7eead6093\") " pod="openshift-infra/auto-csr-approver-29556468-njzx8" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.469215 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556468-njzx8" Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.687990 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556468-njzx8"] Mar 13 07:48:00 crc kubenswrapper[4876]: W0313 07:48:00.700768 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod837c43f4_e275_4668_b5c2_34f7eead6093.slice/crio-a3bfdcf49e52d8d61d9a0021d73c923eda414ac18f20839860fd67928418d98e WatchSource:0}: Error finding container a3bfdcf49e52d8d61d9a0021d73c923eda414ac18f20839860fd67928418d98e: Status 404 returned error can't find the container with id a3bfdcf49e52d8d61d9a0021d73c923eda414ac18f20839860fd67928418d98e Mar 13 07:48:00 crc kubenswrapper[4876]: I0313 07:48:00.704297 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 07:48:01 crc kubenswrapper[4876]: I0313 07:48:01.513905 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556468-njzx8" event={"ID":"837c43f4-e275-4668-b5c2-34f7eead6093","Type":"ContainerStarted","Data":"a3bfdcf49e52d8d61d9a0021d73c923eda414ac18f20839860fd67928418d98e"} Mar 13 07:48:02 crc kubenswrapper[4876]: I0313 07:48:02.523480 4876 generic.go:334] "Generic (PLEG): container finished" podID="837c43f4-e275-4668-b5c2-34f7eead6093" containerID="1f45548d006a179e52d206bfa29a1c4a272b10d2ec033b3ccd7d01bb5796082a" exitCode=0 Mar 13 07:48:02 crc kubenswrapper[4876]: I0313 07:48:02.523556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556468-njzx8" event={"ID":"837c43f4-e275-4668-b5c2-34f7eead6093","Type":"ContainerDied","Data":"1f45548d006a179e52d206bfa29a1c4a272b10d2ec033b3ccd7d01bb5796082a"} Mar 13 07:48:03 crc kubenswrapper[4876]: I0313 07:48:03.821662 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556468-njzx8" Mar 13 07:48:03 crc kubenswrapper[4876]: I0313 07:48:03.939802 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt8qj\" (UniqueName: \"kubernetes.io/projected/837c43f4-e275-4668-b5c2-34f7eead6093-kube-api-access-wt8qj\") pod \"837c43f4-e275-4668-b5c2-34f7eead6093\" (UID: \"837c43f4-e275-4668-b5c2-34f7eead6093\") " Mar 13 07:48:03 crc kubenswrapper[4876]: I0313 07:48:03.949647 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/837c43f4-e275-4668-b5c2-34f7eead6093-kube-api-access-wt8qj" (OuterVolumeSpecName: "kube-api-access-wt8qj") pod "837c43f4-e275-4668-b5c2-34f7eead6093" (UID: "837c43f4-e275-4668-b5c2-34f7eead6093"). InnerVolumeSpecName "kube-api-access-wt8qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:48:04 crc kubenswrapper[4876]: I0313 07:48:04.041786 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt8qj\" (UniqueName: \"kubernetes.io/projected/837c43f4-e275-4668-b5c2-34f7eead6093-kube-api-access-wt8qj\") on node \"crc\" DevicePath \"\"" Mar 13 07:48:04 crc kubenswrapper[4876]: I0313 07:48:04.539004 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556468-njzx8" event={"ID":"837c43f4-e275-4668-b5c2-34f7eead6093","Type":"ContainerDied","Data":"a3bfdcf49e52d8d61d9a0021d73c923eda414ac18f20839860fd67928418d98e"} Mar 13 07:48:04 crc kubenswrapper[4876]: I0313 07:48:04.539053 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556468-njzx8" Mar 13 07:48:04 crc kubenswrapper[4876]: I0313 07:48:04.539056 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3bfdcf49e52d8d61d9a0021d73c923eda414ac18f20839860fd67928418d98e" Mar 13 07:48:04 crc kubenswrapper[4876]: I0313 07:48:04.885467 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556462-xxthz"] Mar 13 07:48:04 crc kubenswrapper[4876]: I0313 07:48:04.888834 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556462-xxthz"] Mar 13 07:48:05 crc kubenswrapper[4876]: I0313 07:48:05.041371 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e125dd-87e2-4362-842e-849288698994" path="/var/lib/kubelet/pods/01e125dd-87e2-4362-842e-849288698994/volumes" Mar 13 07:49:24 crc kubenswrapper[4876]: I0313 07:49:24.607954 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:49:24 crc kubenswrapper[4876]: I0313 07:49:24.609276 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:49:54 crc kubenswrapper[4876]: I0313 07:49:54.607953 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:49:54 crc kubenswrapper[4876]: I0313 07:49:54.609100 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.146179 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556470-k8djl"] Mar 13 07:50:00 crc kubenswrapper[4876]: E0313 07:50:00.147583 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="837c43f4-e275-4668-b5c2-34f7eead6093" containerName="oc" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.147617 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="837c43f4-e275-4668-b5c2-34f7eead6093" containerName="oc" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.147802 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="837c43f4-e275-4668-b5c2-34f7eead6093" containerName="oc" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.148518 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556470-k8djl" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.151568 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556470-k8djl"] Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.152542 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.152622 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.152760 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.291507 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wwnj\" (UniqueName: \"kubernetes.io/projected/34633e40-856c-4f46-a981-1cb41f46dacd-kube-api-access-5wwnj\") pod \"auto-csr-approver-29556470-k8djl\" (UID: \"34633e40-856c-4f46-a981-1cb41f46dacd\") " pod="openshift-infra/auto-csr-approver-29556470-k8djl" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.393078 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wwnj\" (UniqueName: \"kubernetes.io/projected/34633e40-856c-4f46-a981-1cb41f46dacd-kube-api-access-5wwnj\") pod \"auto-csr-approver-29556470-k8djl\" (UID: \"34633e40-856c-4f46-a981-1cb41f46dacd\") " pod="openshift-infra/auto-csr-approver-29556470-k8djl" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.425945 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wwnj\" (UniqueName: \"kubernetes.io/projected/34633e40-856c-4f46-a981-1cb41f46dacd-kube-api-access-5wwnj\") pod \"auto-csr-approver-29556470-k8djl\" (UID: \"34633e40-856c-4f46-a981-1cb41f46dacd\") " pod="openshift-infra/auto-csr-approver-29556470-k8djl" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.477510 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556470-k8djl" Mar 13 07:50:00 crc kubenswrapper[4876]: I0313 07:50:00.671459 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556470-k8djl"] Mar 13 07:50:01 crc kubenswrapper[4876]: I0313 07:50:01.319985 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556470-k8djl" event={"ID":"34633e40-856c-4f46-a981-1cb41f46dacd","Type":"ContainerStarted","Data":"f7722d4205a055db2ec4ebc81914af0930d5439a7da555fc361456d5183857fa"} Mar 13 07:50:01 crc kubenswrapper[4876]: I0313 07:50:01.388726 4876 scope.go:117] "RemoveContainer" containerID="dd88861adab01940a851ebcd8cbefba44db0e5bbadef910dded117593827a74f" Mar 13 07:50:01 crc kubenswrapper[4876]: I0313 07:50:01.408388 4876 scope.go:117] "RemoveContainer" containerID="5a8c1853fd57f8e91f5ac9a13a6a8e9950f23601731d0d4b9947ae2713569eed" Mar 13 07:50:01 crc kubenswrapper[4876]: I0313 07:50:01.443409 4876 scope.go:117] "RemoveContainer" containerID="3fd17f9a4ded5706d5a26792a484912481f46bc41b5c9854d5f61914d6af7696" Mar 13 07:50:02 crc kubenswrapper[4876]: I0313 07:50:02.327541 4876 generic.go:334] "Generic (PLEG): container finished" podID="34633e40-856c-4f46-a981-1cb41f46dacd" containerID="89186f77092dbe556c0d7aa5a6c5c92db61dc352f103ab14880e73967e39e98c" exitCode=0 Mar 13 07:50:02 crc kubenswrapper[4876]: I0313 07:50:02.327598 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556470-k8djl" event={"ID":"34633e40-856c-4f46-a981-1cb41f46dacd","Type":"ContainerDied","Data":"89186f77092dbe556c0d7aa5a6c5c92db61dc352f103ab14880e73967e39e98c"} Mar 13 07:50:03 crc kubenswrapper[4876]: I0313 07:50:03.531515 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556470-k8djl" Mar 13 07:50:03 crc kubenswrapper[4876]: I0313 07:50:03.635541 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wwnj\" (UniqueName: \"kubernetes.io/projected/34633e40-856c-4f46-a981-1cb41f46dacd-kube-api-access-5wwnj\") pod \"34633e40-856c-4f46-a981-1cb41f46dacd\" (UID: \"34633e40-856c-4f46-a981-1cb41f46dacd\") " Mar 13 07:50:03 crc kubenswrapper[4876]: I0313 07:50:03.641578 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34633e40-856c-4f46-a981-1cb41f46dacd-kube-api-access-5wwnj" (OuterVolumeSpecName: "kube-api-access-5wwnj") pod "34633e40-856c-4f46-a981-1cb41f46dacd" (UID: "34633e40-856c-4f46-a981-1cb41f46dacd"). InnerVolumeSpecName "kube-api-access-5wwnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:50:03 crc kubenswrapper[4876]: I0313 07:50:03.737393 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wwnj\" (UniqueName: \"kubernetes.io/projected/34633e40-856c-4f46-a981-1cb41f46dacd-kube-api-access-5wwnj\") on node \"crc\" DevicePath \"\"" Mar 13 07:50:04 crc kubenswrapper[4876]: I0313 07:50:04.342584 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556470-k8djl" event={"ID":"34633e40-856c-4f46-a981-1cb41f46dacd","Type":"ContainerDied","Data":"f7722d4205a055db2ec4ebc81914af0930d5439a7da555fc361456d5183857fa"} Mar 13 07:50:04 crc kubenswrapper[4876]: I0313 07:50:04.342630 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7722d4205a055db2ec4ebc81914af0930d5439a7da555fc361456d5183857fa" Mar 13 07:50:04 crc kubenswrapper[4876]: I0313 07:50:04.342655 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556470-k8djl" Mar 13 07:50:04 crc kubenswrapper[4876]: I0313 07:50:04.597539 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556464-ctthh"] Mar 13 07:50:04 crc kubenswrapper[4876]: I0313 07:50:04.601838 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556464-ctthh"] Mar 13 07:50:05 crc kubenswrapper[4876]: I0313 07:50:05.049454 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="134014ac-2aa1-411c-a5af-492088dee314" path="/var/lib/kubelet/pods/134014ac-2aa1-411c-a5af-492088dee314/volumes" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.631376 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k2mgm"] Mar 13 07:50:13 crc kubenswrapper[4876]: E0313 07:50:13.632631 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34633e40-856c-4f46-a981-1cb41f46dacd" containerName="oc" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.632650 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="34633e40-856c-4f46-a981-1cb41f46dacd" containerName="oc" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.632772 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="34633e40-856c-4f46-a981-1cb41f46dacd" containerName="oc" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.633339 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.649393 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k2mgm"] Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.801666 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.801794 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-registry-certificates\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.801835 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-trusted-ca\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.801875 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.801918 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jfgk\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-kube-api-access-6jfgk\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.801976 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-bound-sa-token\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.802260 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.802343 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-registry-tls\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.833538 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.903721 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jfgk\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-kube-api-access-6jfgk\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.903801 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-bound-sa-token\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.903890 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.903932 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-registry-tls\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.904058 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-registry-certificates\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.904097 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-trusted-ca\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.904140 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.905396 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.906142 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-registry-certificates\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.906271 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-trusted-ca\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.913305 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-registry-tls\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.914853 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.936584 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-bound-sa-token\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.941980 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jfgk\" (UniqueName: \"kubernetes.io/projected/9ac5011a-806e-4024-9fd9-8e1f9550e4a1-kube-api-access-6jfgk\") pod \"image-registry-66df7c8f76-k2mgm\" (UID: \"9ac5011a-806e-4024-9fd9-8e1f9550e4a1\") " pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:13 crc kubenswrapper[4876]: I0313 07:50:13.951708 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:14 crc kubenswrapper[4876]: I0313 07:50:14.292945 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k2mgm"] Mar 13 07:50:14 crc kubenswrapper[4876]: I0313 07:50:14.415798 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" event={"ID":"9ac5011a-806e-4024-9fd9-8e1f9550e4a1","Type":"ContainerStarted","Data":"c2acc52e35dbb5b19adcb558528c9b7ffccf44b1948f4c46de940e1d87ffdc64"} Mar 13 07:50:15 crc kubenswrapper[4876]: I0313 07:50:15.427675 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" event={"ID":"9ac5011a-806e-4024-9fd9-8e1f9550e4a1","Type":"ContainerStarted","Data":"bf5784239ce7b147f6d37f5c9c4512ccf7fc93e17a3d43e14a10eb29f98677fb"} Mar 13 07:50:15 crc kubenswrapper[4876]: I0313 07:50:15.428192 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:15 crc kubenswrapper[4876]: I0313 07:50:15.464138 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" podStartSLOduration=2.464117335 podStartE2EDuration="2.464117335s" podCreationTimestamp="2026-03-13 07:50:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:50:15.461890632 +0000 UTC m=+675.132669624" watchObservedRunningTime="2026-03-13 07:50:15.464117335 +0000 UTC m=+675.134896327" Mar 13 07:50:24 crc kubenswrapper[4876]: I0313 07:50:24.608593 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:50:24 crc kubenswrapper[4876]: I0313 07:50:24.609505 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:50:24 crc kubenswrapper[4876]: I0313 07:50:24.609578 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:50:24 crc kubenswrapper[4876]: I0313 07:50:24.610512 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba651ad626e0def50adb9d0b266fb9ee781c393b76906dacbf80c73b1a02e496"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 07:50:24 crc kubenswrapper[4876]: I0313 07:50:24.610593 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://ba651ad626e0def50adb9d0b266fb9ee781c393b76906dacbf80c73b1a02e496" gracePeriod=600 Mar 13 07:50:25 crc kubenswrapper[4876]: I0313 07:50:25.522522 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="ba651ad626e0def50adb9d0b266fb9ee781c393b76906dacbf80c73b1a02e496" exitCode=0 Mar 13 07:50:25 crc kubenswrapper[4876]: I0313 07:50:25.522563 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"ba651ad626e0def50adb9d0b266fb9ee781c393b76906dacbf80c73b1a02e496"} Mar 13 07:50:25 crc kubenswrapper[4876]: I0313 07:50:25.523373 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"8ebde89ffdd3ea68ea7599ed04541e15278a60938a622659662a7eee8835605e"} Mar 13 07:50:25 crc kubenswrapper[4876]: I0313 07:50:25.523406 4876 scope.go:117] "RemoveContainer" containerID="9f60a1005aa99b7a7a6f69d8ddd25eddf167d10b844bb62bc7097f7905502e52" Mar 13 07:50:33 crc kubenswrapper[4876]: I0313 07:50:33.959179 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-k2mgm" Mar 13 07:50:34 crc kubenswrapper[4876]: I0313 07:50:34.067004 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5pvl"] Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.109454 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" podUID="0991deda-4e1a-4fe7-92e9-01bbaf54efb0" containerName="registry" containerID="cri-o://2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a" gracePeriod=30 Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.532076 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.619141 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-ca-trust-extracted\") pod \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.619410 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-installation-pull-secrets\") pod \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.619515 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ldtm\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-kube-api-access-5ldtm\") pod \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.619604 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-tls\") pod \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.619679 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-certificates\") pod \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.619830 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-bound-sa-token\") pod \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.620135 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.620379 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-trusted-ca\") pod \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\" (UID: \"0991deda-4e1a-4fe7-92e9-01bbaf54efb0\") " Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.622211 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "0991deda-4e1a-4fe7-92e9-01bbaf54efb0" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.622527 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "0991deda-4e1a-4fe7-92e9-01bbaf54efb0" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.625864 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-kube-api-access-5ldtm" (OuterVolumeSpecName: "kube-api-access-5ldtm") pod "0991deda-4e1a-4fe7-92e9-01bbaf54efb0" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0"). InnerVolumeSpecName "kube-api-access-5ldtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.625996 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "0991deda-4e1a-4fe7-92e9-01bbaf54efb0" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.626811 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "0991deda-4e1a-4fe7-92e9-01bbaf54efb0" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.627001 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "0991deda-4e1a-4fe7-92e9-01bbaf54efb0" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.633385 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "0991deda-4e1a-4fe7-92e9-01bbaf54efb0" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.644052 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "0991deda-4e1a-4fe7-92e9-01bbaf54efb0" (UID: "0991deda-4e1a-4fe7-92e9-01bbaf54efb0"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.722193 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.722298 4876 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.722320 4876 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.722342 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ldtm\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-kube-api-access-5ldtm\") on node \"crc\" DevicePath \"\"" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.722361 4876 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.722379 4876 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.722396 4876 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0991deda-4e1a-4fe7-92e9-01bbaf54efb0-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.795809 4876 generic.go:334] "Generic (PLEG): container finished" podID="0991deda-4e1a-4fe7-92e9-01bbaf54efb0" containerID="2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a" exitCode=0 Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.795869 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" event={"ID":"0991deda-4e1a-4fe7-92e9-01bbaf54efb0","Type":"ContainerDied","Data":"2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a"} Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.795904 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" event={"ID":"0991deda-4e1a-4fe7-92e9-01bbaf54efb0","Type":"ContainerDied","Data":"25060b51be92f6e5fb37c755cd1c6aebc9008c6acca9803c25642996b49d3e5a"} Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.795921 4876 scope.go:117] "RemoveContainer" containerID="2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.795922 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c5pvl" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.831069 4876 scope.go:117] "RemoveContainer" containerID="2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a" Mar 13 07:50:59 crc kubenswrapper[4876]: E0313 07:50:59.834087 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a\": container with ID starting with 2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a not found: ID does not exist" containerID="2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.835089 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a"} err="failed to get container status \"2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a\": rpc error: code = NotFound desc = could not find container \"2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a\": container with ID starting with 2178f95adcd04c3a0c76c0fd1dcc3e26d631c6e6d8cf61b1c7d51a4bcd68f98a not found: ID does not exist" Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.842495 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5pvl"] Mar 13 07:50:59 crc kubenswrapper[4876]: I0313 07:50:59.850065 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c5pvl"] Mar 13 07:51:01 crc kubenswrapper[4876]: I0313 07:51:01.042285 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0991deda-4e1a-4fe7-92e9-01bbaf54efb0" path="/var/lib/kubelet/pods/0991deda-4e1a-4fe7-92e9-01bbaf54efb0/volumes" Mar 13 07:51:01 crc kubenswrapper[4876]: I0313 07:51:01.489879 4876 scope.go:117] "RemoveContainer" containerID="e98d759f76ecebdc8f337c52e8649cb616f630db684e7ba0690a4b677228d42e" Mar 13 07:51:01 crc kubenswrapper[4876]: I0313 07:51:01.545969 4876 scope.go:117] "RemoveContainer" containerID="ebe01a9218f14fa2b963c4a3e694ae49107fac9412b157ec4178fe1304bbe3d3" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.334064 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw"] Mar 13 07:51:18 crc kubenswrapper[4876]: E0313 07:51:18.335518 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0991deda-4e1a-4fe7-92e9-01bbaf54efb0" containerName="registry" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.335544 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0991deda-4e1a-4fe7-92e9-01bbaf54efb0" containerName="registry" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.335702 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0991deda-4e1a-4fe7-92e9-01bbaf54efb0" containerName="registry" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.336352 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.338301 4876 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-d4vxs" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.338929 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.339122 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.343656 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-552r7"] Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.344592 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-552r7" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.349286 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw"] Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.356840 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-jm7bb"] Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.356895 4876 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-kssn2" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.357819 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.359641 4876 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-595qt" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.363252 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-552r7"] Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.369500 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-jm7bb"] Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.404379 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75f24\" (UniqueName: \"kubernetes.io/projected/d0aa3593-da37-4da6-ae1f-890cb901c029-kube-api-access-75f24\") pod \"cert-manager-webhook-687f57d79b-jm7bb\" (UID: \"d0aa3593-da37-4da6-ae1f-890cb901c029\") " pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.404489 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdl52\" (UniqueName: \"kubernetes.io/projected/bbd6238b-a711-47f6-bf54-ae2b35e8c79d-kube-api-access-bdl52\") pod \"cert-manager-858654f9db-552r7\" (UID: \"bbd6238b-a711-47f6-bf54-ae2b35e8c79d\") " pod="cert-manager/cert-manager-858654f9db-552r7" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.404521 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq9n8\" (UniqueName: \"kubernetes.io/projected/037a0101-dff7-49a6-9974-a5b88c2cb7f1-kube-api-access-rq9n8\") pod \"cert-manager-cainjector-cf98fcc89-2c8kw\" (UID: \"037a0101-dff7-49a6-9974-a5b88c2cb7f1\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.506180 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdl52\" (UniqueName: \"kubernetes.io/projected/bbd6238b-a711-47f6-bf54-ae2b35e8c79d-kube-api-access-bdl52\") pod \"cert-manager-858654f9db-552r7\" (UID: \"bbd6238b-a711-47f6-bf54-ae2b35e8c79d\") " pod="cert-manager/cert-manager-858654f9db-552r7" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.506254 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq9n8\" (UniqueName: \"kubernetes.io/projected/037a0101-dff7-49a6-9974-a5b88c2cb7f1-kube-api-access-rq9n8\") pod \"cert-manager-cainjector-cf98fcc89-2c8kw\" (UID: \"037a0101-dff7-49a6-9974-a5b88c2cb7f1\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.506306 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75f24\" (UniqueName: \"kubernetes.io/projected/d0aa3593-da37-4da6-ae1f-890cb901c029-kube-api-access-75f24\") pod \"cert-manager-webhook-687f57d79b-jm7bb\" (UID: \"d0aa3593-da37-4da6-ae1f-890cb901c029\") " pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.529922 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq9n8\" (UniqueName: \"kubernetes.io/projected/037a0101-dff7-49a6-9974-a5b88c2cb7f1-kube-api-access-rq9n8\") pod \"cert-manager-cainjector-cf98fcc89-2c8kw\" (UID: \"037a0101-dff7-49a6-9974-a5b88c2cb7f1\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.530109 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdl52\" (UniqueName: \"kubernetes.io/projected/bbd6238b-a711-47f6-bf54-ae2b35e8c79d-kube-api-access-bdl52\") pod \"cert-manager-858654f9db-552r7\" (UID: \"bbd6238b-a711-47f6-bf54-ae2b35e8c79d\") " pod="cert-manager/cert-manager-858654f9db-552r7" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.530891 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75f24\" (UniqueName: \"kubernetes.io/projected/d0aa3593-da37-4da6-ae1f-890cb901c029-kube-api-access-75f24\") pod \"cert-manager-webhook-687f57d79b-jm7bb\" (UID: \"d0aa3593-da37-4da6-ae1f-890cb901c029\") " pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.662621 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.687692 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-552r7" Mar 13 07:51:18 crc kubenswrapper[4876]: I0313 07:51:18.697169 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" Mar 13 07:51:19 crc kubenswrapper[4876]: I0313 07:51:19.147086 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw"] Mar 13 07:51:19 crc kubenswrapper[4876]: I0313 07:51:19.205671 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-jm7bb"] Mar 13 07:51:19 crc kubenswrapper[4876]: I0313 07:51:19.214941 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-552r7"] Mar 13 07:51:19 crc kubenswrapper[4876]: W0313 07:51:19.218512 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0aa3593_da37_4da6_ae1f_890cb901c029.slice/crio-cab0546f65579e943e74e3e14069ebc43f7b7e2a86f1d5c40167e949562d9eff WatchSource:0}: Error finding container cab0546f65579e943e74e3e14069ebc43f7b7e2a86f1d5c40167e949562d9eff: Status 404 returned error can't find the container with id cab0546f65579e943e74e3e14069ebc43f7b7e2a86f1d5c40167e949562d9eff Mar 13 07:51:19 crc kubenswrapper[4876]: W0313 07:51:19.220650 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbd6238b_a711_47f6_bf54_ae2b35e8c79d.slice/crio-b6bb6491db46b23b07d4e8b8e23fe67ae7050cc14eb68fc038b54a83bfb8e289 WatchSource:0}: Error finding container b6bb6491db46b23b07d4e8b8e23fe67ae7050cc14eb68fc038b54a83bfb8e289: Status 404 returned error can't find the container with id b6bb6491db46b23b07d4e8b8e23fe67ae7050cc14eb68fc038b54a83bfb8e289 Mar 13 07:51:19 crc kubenswrapper[4876]: I0313 07:51:19.944445 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw" event={"ID":"037a0101-dff7-49a6-9974-a5b88c2cb7f1","Type":"ContainerStarted","Data":"2e2c8d3ace347b860e433e3d81b95dfbc51dd5cff9133eeb25c2340f9c2e9610"} Mar 13 07:51:19 crc kubenswrapper[4876]: I0313 07:51:19.946616 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" event={"ID":"d0aa3593-da37-4da6-ae1f-890cb901c029","Type":"ContainerStarted","Data":"cab0546f65579e943e74e3e14069ebc43f7b7e2a86f1d5c40167e949562d9eff"} Mar 13 07:51:19 crc kubenswrapper[4876]: I0313 07:51:19.948025 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-552r7" event={"ID":"bbd6238b-a711-47f6-bf54-ae2b35e8c79d","Type":"ContainerStarted","Data":"b6bb6491db46b23b07d4e8b8e23fe67ae7050cc14eb68fc038b54a83bfb8e289"} Mar 13 07:51:21 crc kubenswrapper[4876]: I0313 07:51:21.964196 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-552r7" event={"ID":"bbd6238b-a711-47f6-bf54-ae2b35e8c79d","Type":"ContainerStarted","Data":"d0976ff310ac4cde490bcb57d8ec9a6782eaae94866584f5e5fb56df16b35180"} Mar 13 07:51:21 crc kubenswrapper[4876]: I0313 07:51:21.986551 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-552r7" podStartSLOduration=1.671683384 podStartE2EDuration="3.98652559s" podCreationTimestamp="2026-03-13 07:51:18 +0000 UTC" firstStartedPulling="2026-03-13 07:51:19.225876315 +0000 UTC m=+738.896655297" lastFinishedPulling="2026-03-13 07:51:21.540718521 +0000 UTC m=+741.211497503" observedRunningTime="2026-03-13 07:51:21.980944821 +0000 UTC m=+741.651723813" watchObservedRunningTime="2026-03-13 07:51:21.98652559 +0000 UTC m=+741.657304572" Mar 13 07:51:23 crc kubenswrapper[4876]: I0313 07:51:23.981321 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw" event={"ID":"037a0101-dff7-49a6-9974-a5b88c2cb7f1","Type":"ContainerStarted","Data":"aa6ffa0e6aa0f708a9373f7238fba60160cb5b45d1b0e5d76197148f6b0310f3"} Mar 13 07:51:23 crc kubenswrapper[4876]: I0313 07:51:23.983119 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" event={"ID":"d0aa3593-da37-4da6-ae1f-890cb901c029","Type":"ContainerStarted","Data":"5b0924fb12c497bd84ac6d23621701fb44a203e36600cddf455295d6378fd2da"} Mar 13 07:51:23 crc kubenswrapper[4876]: I0313 07:51:23.983302 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" Mar 13 07:51:24 crc kubenswrapper[4876]: I0313 07:51:24.000475 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2c8kw" podStartSLOduration=2.173670213 podStartE2EDuration="6.000435527s" podCreationTimestamp="2026-03-13 07:51:18 +0000 UTC" firstStartedPulling="2026-03-13 07:51:19.155959432 +0000 UTC m=+738.826738414" lastFinishedPulling="2026-03-13 07:51:22.982724746 +0000 UTC m=+742.653503728" observedRunningTime="2026-03-13 07:51:23.997120563 +0000 UTC m=+743.667899555" watchObservedRunningTime="2026-03-13 07:51:24.000435527 +0000 UTC m=+743.671214539" Mar 13 07:51:24 crc kubenswrapper[4876]: I0313 07:51:24.021411 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" podStartSLOduration=2.2200293540000002 podStartE2EDuration="6.021379954s" podCreationTimestamp="2026-03-13 07:51:18 +0000 UTC" firstStartedPulling="2026-03-13 07:51:19.220843101 +0000 UTC m=+738.891622083" lastFinishedPulling="2026-03-13 07:51:23.022193701 +0000 UTC m=+742.692972683" observedRunningTime="2026-03-13 07:51:24.016935778 +0000 UTC m=+743.687714790" watchObservedRunningTime="2026-03-13 07:51:24.021379954 +0000 UTC m=+743.692158956" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.322363 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-g4xkn"] Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.324656 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovn-controller" containerID="cri-o://93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff" gracePeriod=30 Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.324741 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="nbdb" containerID="cri-o://01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf" gracePeriod=30 Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.324901 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="northd" containerID="cri-o://2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0" gracePeriod=30 Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.325030 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="sbdb" containerID="cri-o://2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9" gracePeriod=30 Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.324931 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899" gracePeriod=30 Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.324984 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovn-acl-logging" containerID="cri-o://3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305" gracePeriod=30 Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.324965 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kube-rbac-proxy-node" containerID="cri-o://72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6" gracePeriod=30 Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.384522 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" containerID="cri-o://7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789" gracePeriod=30 Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.668443 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/3.log" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.670910 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovn-acl-logging/0.log" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.671547 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovn-controller/0.log" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.672197 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.701342 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-jm7bb" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752483 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9bxnt"] Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752798 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752815 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752827 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752836 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752853 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752861 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752872 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752880 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752894 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752902 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752913 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kube-rbac-proxy-node" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752923 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kube-rbac-proxy-node" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752940 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="nbdb" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752950 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="nbdb" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752960 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="sbdb" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752968 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="sbdb" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752978 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovn-acl-logging" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.752986 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovn-acl-logging" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.752995 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kubecfg-setup" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753004 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kubecfg-setup" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.753017 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovn-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753026 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovn-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.753041 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="northd" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753049 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="northd" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753168 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753179 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovn-acl-logging" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753193 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753205 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753214 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovn-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753223 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="northd" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753253 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="nbdb" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753266 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753277 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="kube-rbac-proxy-node" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753287 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="sbdb" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753300 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: E0313 07:51:28.753428 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753439 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.753563 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerName="ovnkube-controller" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.755741 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.757805 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.757876 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-kubelet\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.757877 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.757900 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-systemd-units\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.757928 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.757947 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-node-log\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.757959 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.757987 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-var-lib-openvswitch\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758005 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-node-log" (OuterVolumeSpecName: "node-log") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758021 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-env-overrides\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758042 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-bin\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758052 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758091 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovn-node-metrics-cert\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758114 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-openvswitch\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758139 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-script-lib\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758156 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-etc-openvswitch\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758189 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-netns\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758214 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-systemd\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758282 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-ovn\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758307 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-config\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758331 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-netd\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758358 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-log-socket\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758394 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-slash\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758411 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758416 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-ovn-kubernetes\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758488 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s45rv\" (UniqueName: \"kubernetes.io/projected/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-kube-api-access-s45rv\") pod \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\" (UID: \"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd\") " Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758591 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-log-socket\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758623 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-ovn\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758648 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-var-lib-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758672 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-ovnkube-config\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758713 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-etc-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758754 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9chh\" (UniqueName: \"kubernetes.io/projected/f22530c5-04ed-4284-bbff-61bb844d8c40-kube-api-access-t9chh\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758775 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-systemd\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758804 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-env-overrides\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758841 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758864 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-cni-bin\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758884 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-slash\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758904 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-run-netns\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758933 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-cni-netd\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758983 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-kubelet\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759010 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-run-ovn-kubernetes\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759032 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f22530c5-04ed-4284-bbff-61bb844d8c40-ovn-node-metrics-cert\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759068 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-systemd-units\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759089 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-ovnkube-script-lib\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759118 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759145 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-node-log\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759193 4876 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759208 4876 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759224 4876 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759262 4876 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759277 4876 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759290 4876 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-node-log\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758488 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.761773 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.758511 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.759253 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.761420 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.761537 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.761588 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.761959 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.761751 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-log-socket" (OuterVolumeSpecName: "log-socket") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.761835 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-slash" (OuterVolumeSpecName: "host-slash") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.762161 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.767845 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-kube-api-access-s45rv" (OuterVolumeSpecName: "kube-api-access-s45rv") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "kube-api-access-s45rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.768351 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.776400 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" (UID: "6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861048 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-ovn\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861152 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-var-lib-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861192 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-ovnkube-config\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861279 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-etc-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861338 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9chh\" (UniqueName: \"kubernetes.io/projected/f22530c5-04ed-4284-bbff-61bb844d8c40-kube-api-access-t9chh\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861372 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-systemd\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861427 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-env-overrides\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861485 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861517 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-cni-bin\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861548 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-slash\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-run-netns\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861623 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-cni-netd\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861669 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-kubelet\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861705 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-run-ovn-kubernetes\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861727 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-cni-bin\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861800 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-ovn\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861833 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861733 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-run-systemd\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861888 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-kubelet\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861889 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-cni-netd\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861920 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-slash\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861735 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f22530c5-04ed-4284-bbff-61bb844d8c40-ovn-node-metrics-cert\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861971 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-run-netns\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861976 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-run-ovn-kubernetes\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861883 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-var-lib-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.861930 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-etc-openvswitch\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862101 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-systemd-units\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862126 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-ovnkube-script-lib\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862162 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862200 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-node-log\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862265 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862277 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-log-socket\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862308 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-log-socket\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862214 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-systemd-units\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862332 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f22530c5-04ed-4284-bbff-61bb844d8c40-node-log\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862361 4876 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862380 4876 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862391 4876 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862401 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862413 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862423 4876 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862432 4876 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862445 4876 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862454 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862463 4876 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862474 4876 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-log-socket\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862487 4876 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-slash\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862503 4876 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.862516 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s45rv\" (UniqueName: \"kubernetes.io/projected/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd-kube-api-access-s45rv\") on node \"crc\" DevicePath \"\"" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.863129 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-env-overrides\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.863319 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-ovnkube-config\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.863554 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f22530c5-04ed-4284-bbff-61bb844d8c40-ovnkube-script-lib\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.866628 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f22530c5-04ed-4284-bbff-61bb844d8c40-ovn-node-metrics-cert\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:28 crc kubenswrapper[4876]: I0313 07:51:28.885334 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9chh\" (UniqueName: \"kubernetes.io/projected/f22530c5-04ed-4284-bbff-61bb844d8c40-kube-api-access-t9chh\") pod \"ovnkube-node-9bxnt\" (UID: \"f22530c5-04ed-4284-bbff-61bb844d8c40\") " pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.020635 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/2.log" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.021450 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/1.log" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.021546 4876 generic.go:334] "Generic (PLEG): container finished" podID="8adb2a69-d0ef-4efc-813d-77fdf9713d50" containerID="76512308b59931886eee75e5bd8c0ff26eed1bfe288f2ff41e80a8ca4761ad73" exitCode=2 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.021631 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-st667" event={"ID":"8adb2a69-d0ef-4efc-813d-77fdf9713d50","Type":"ContainerDied","Data":"76512308b59931886eee75e5bd8c0ff26eed1bfe288f2ff41e80a8ca4761ad73"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.021840 4876 scope.go:117] "RemoveContainer" containerID="a62ff4c03335fae410c16f3b745d90a193e67eff5e762ffa140d2ba2f4384eb4" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.022771 4876 scope.go:117] "RemoveContainer" containerID="76512308b59931886eee75e5bd8c0ff26eed1bfe288f2ff41e80a8ca4761ad73" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.023413 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-st667_openshift-multus(8adb2a69-d0ef-4efc-813d-77fdf9713d50)\"" pod="openshift-multus/multus-st667" podUID="8adb2a69-d0ef-4efc-813d-77fdf9713d50" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.024779 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovnkube-controller/3.log" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.028500 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovn-acl-logging/0.log" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029135 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-g4xkn_6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/ovn-controller/0.log" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029588 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789" exitCode=0 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029637 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9" exitCode=0 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029658 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf" exitCode=0 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029716 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0" exitCode=0 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029731 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899" exitCode=0 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029750 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6" exitCode=0 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029765 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305" exitCode=143 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029784 4876 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" containerID="93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff" exitCode=143 Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029762 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029856 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029888 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029912 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029930 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029946 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029962 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.029980 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030005 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030017 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030027 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030037 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030049 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030058 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030066 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030075 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030082 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030094 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030113 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030122 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030129 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030137 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030145 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030153 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030160 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030168 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030175 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030182 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030193 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030206 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030214 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030222 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030229 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030267 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030275 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030282 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030289 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030298 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030316 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030333 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g4xkn" event={"ID":"6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd","Type":"ContainerDied","Data":"0e9da90db3918b231f3e15e05b6b39eab6b349dc5e7ff7cbb9584c79f673692e"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030352 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030365 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030374 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030384 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030393 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030402 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030411 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030491 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030504 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.030514 4876 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.072547 4876 scope.go:117] "RemoveContainer" containerID="7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.075227 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.100163 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-g4xkn"] Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.108508 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.111487 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-g4xkn"] Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.130687 4876 scope.go:117] "RemoveContainer" containerID="2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.152307 4876 scope.go:117] "RemoveContainer" containerID="01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.172356 4876 scope.go:117] "RemoveContainer" containerID="2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.194776 4876 scope.go:117] "RemoveContainer" containerID="69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.210921 4876 scope.go:117] "RemoveContainer" containerID="72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.229313 4876 scope.go:117] "RemoveContainer" containerID="3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.298337 4876 scope.go:117] "RemoveContainer" containerID="93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.316811 4876 scope.go:117] "RemoveContainer" containerID="1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.332399 4876 scope.go:117] "RemoveContainer" containerID="7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.332793 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": container with ID starting with 7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789 not found: ID does not exist" containerID="7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.332828 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} err="failed to get container status \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": rpc error: code = NotFound desc = could not find container \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": container with ID starting with 7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.332856 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.333172 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": container with ID starting with 336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3 not found: ID does not exist" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.333220 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} err="failed to get container status \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": rpc error: code = NotFound desc = could not find container \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": container with ID starting with 336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.333271 4876 scope.go:117] "RemoveContainer" containerID="2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.333734 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": container with ID starting with 2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9 not found: ID does not exist" containerID="2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.333763 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} err="failed to get container status \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": rpc error: code = NotFound desc = could not find container \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": container with ID starting with 2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.333781 4876 scope.go:117] "RemoveContainer" containerID="01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.334149 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": container with ID starting with 01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf not found: ID does not exist" containerID="01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.334180 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} err="failed to get container status \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": rpc error: code = NotFound desc = could not find container \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": container with ID starting with 01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.334201 4876 scope.go:117] "RemoveContainer" containerID="2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.334515 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": container with ID starting with 2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0 not found: ID does not exist" containerID="2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.334554 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} err="failed to get container status \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": rpc error: code = NotFound desc = could not find container \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": container with ID starting with 2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.334575 4876 scope.go:117] "RemoveContainer" containerID="69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.334856 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": container with ID starting with 69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899 not found: ID does not exist" containerID="69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.334893 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} err="failed to get container status \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": rpc error: code = NotFound desc = could not find container \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": container with ID starting with 69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.334915 4876 scope.go:117] "RemoveContainer" containerID="72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.335181 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": container with ID starting with 72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6 not found: ID does not exist" containerID="72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.335212 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} err="failed to get container status \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": rpc error: code = NotFound desc = could not find container \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": container with ID starting with 72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.335230 4876 scope.go:117] "RemoveContainer" containerID="3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.335739 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": container with ID starting with 3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305 not found: ID does not exist" containerID="3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.335770 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} err="failed to get container status \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": rpc error: code = NotFound desc = could not find container \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": container with ID starting with 3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.335790 4876 scope.go:117] "RemoveContainer" containerID="93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.336071 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": container with ID starting with 93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff not found: ID does not exist" containerID="93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.336099 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} err="failed to get container status \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": rpc error: code = NotFound desc = could not find container \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": container with ID starting with 93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.336114 4876 scope.go:117] "RemoveContainer" containerID="1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf" Mar 13 07:51:29 crc kubenswrapper[4876]: E0313 07:51:29.336394 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": container with ID starting with 1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf not found: ID does not exist" containerID="1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.336421 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} err="failed to get container status \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": rpc error: code = NotFound desc = could not find container \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": container with ID starting with 1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.336438 4876 scope.go:117] "RemoveContainer" containerID="7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.336685 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} err="failed to get container status \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": rpc error: code = NotFound desc = could not find container \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": container with ID starting with 7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.336709 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.336984 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} err="failed to get container status \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": rpc error: code = NotFound desc = could not find container \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": container with ID starting with 336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.337031 4876 scope.go:117] "RemoveContainer" containerID="2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.337343 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} err="failed to get container status \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": rpc error: code = NotFound desc = could not find container \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": container with ID starting with 2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.337375 4876 scope.go:117] "RemoveContainer" containerID="01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.337689 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} err="failed to get container status \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": rpc error: code = NotFound desc = could not find container \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": container with ID starting with 01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.337716 4876 scope.go:117] "RemoveContainer" containerID="2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.337988 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} err="failed to get container status \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": rpc error: code = NotFound desc = could not find container \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": container with ID starting with 2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.338016 4876 scope.go:117] "RemoveContainer" containerID="69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.338455 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} err="failed to get container status \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": rpc error: code = NotFound desc = could not find container \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": container with ID starting with 69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.338485 4876 scope.go:117] "RemoveContainer" containerID="72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.338728 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} err="failed to get container status \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": rpc error: code = NotFound desc = could not find container \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": container with ID starting with 72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.338755 4876 scope.go:117] "RemoveContainer" containerID="3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.339084 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} err="failed to get container status \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": rpc error: code = NotFound desc = could not find container \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": container with ID starting with 3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.339110 4876 scope.go:117] "RemoveContainer" containerID="93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.339431 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} err="failed to get container status \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": rpc error: code = NotFound desc = could not find container \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": container with ID starting with 93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.339454 4876 scope.go:117] "RemoveContainer" containerID="1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.339744 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} err="failed to get container status \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": rpc error: code = NotFound desc = could not find container \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": container with ID starting with 1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.339768 4876 scope.go:117] "RemoveContainer" containerID="7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.340040 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} err="failed to get container status \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": rpc error: code = NotFound desc = could not find container \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": container with ID starting with 7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.340078 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.340554 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} err="failed to get container status \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": rpc error: code = NotFound desc = could not find container \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": container with ID starting with 336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.340583 4876 scope.go:117] "RemoveContainer" containerID="2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.340797 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} err="failed to get container status \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": rpc error: code = NotFound desc = could not find container \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": container with ID starting with 2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.340824 4876 scope.go:117] "RemoveContainer" containerID="01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.341105 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} err="failed to get container status \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": rpc error: code = NotFound desc = could not find container \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": container with ID starting with 01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.341131 4876 scope.go:117] "RemoveContainer" containerID="2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.341368 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} err="failed to get container status \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": rpc error: code = NotFound desc = could not find container \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": container with ID starting with 2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.341392 4876 scope.go:117] "RemoveContainer" containerID="69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.341728 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} err="failed to get container status \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": rpc error: code = NotFound desc = could not find container \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": container with ID starting with 69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.341759 4876 scope.go:117] "RemoveContainer" containerID="72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.342067 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} err="failed to get container status \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": rpc error: code = NotFound desc = could not find container \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": container with ID starting with 72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.342101 4876 scope.go:117] "RemoveContainer" containerID="3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.342430 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} err="failed to get container status \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": rpc error: code = NotFound desc = could not find container \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": container with ID starting with 3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.342464 4876 scope.go:117] "RemoveContainer" containerID="93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.342725 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} err="failed to get container status \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": rpc error: code = NotFound desc = could not find container \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": container with ID starting with 93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.342751 4876 scope.go:117] "RemoveContainer" containerID="1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.342980 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} err="failed to get container status \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": rpc error: code = NotFound desc = could not find container \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": container with ID starting with 1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.343008 4876 scope.go:117] "RemoveContainer" containerID="7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.343327 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789"} err="failed to get container status \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": rpc error: code = NotFound desc = could not find container \"7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789\": container with ID starting with 7b3bb57af31fb9e557c5813db517a24751d3d72066fba69d1b8f791074cfd789 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.343361 4876 scope.go:117] "RemoveContainer" containerID="336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.343612 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3"} err="failed to get container status \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": rpc error: code = NotFound desc = could not find container \"336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3\": container with ID starting with 336137a84e651a69265ca56e01b60c1c1c2e7ef7e4aaa4eb126ee3b62168b2e3 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.343646 4876 scope.go:117] "RemoveContainer" containerID="2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.343915 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9"} err="failed to get container status \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": rpc error: code = NotFound desc = could not find container \"2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9\": container with ID starting with 2895d835ad1b9ba9cc04972062f345467b7c34a0f93110b973b3dc6fa10dbfd9 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.343943 4876 scope.go:117] "RemoveContainer" containerID="01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.344187 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf"} err="failed to get container status \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": rpc error: code = NotFound desc = could not find container \"01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf\": container with ID starting with 01696f28472f5d621a41f83bcfd0bf0e43adb6b514eef96525de0aab2563c6bf not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.344216 4876 scope.go:117] "RemoveContainer" containerID="2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.344536 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0"} err="failed to get container status \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": rpc error: code = NotFound desc = could not find container \"2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0\": container with ID starting with 2667bea820b3822336bb5aabc9fada3d64cbe7d2422e755cfd8a868e39b8bfc0 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.344569 4876 scope.go:117] "RemoveContainer" containerID="69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.344809 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899"} err="failed to get container status \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": rpc error: code = NotFound desc = could not find container \"69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899\": container with ID starting with 69997aed3020f695e3a8737a83cd0194f57c6b48700a6c9d5cfb094b45d44899 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.344840 4876 scope.go:117] "RemoveContainer" containerID="72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.345072 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6"} err="failed to get container status \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": rpc error: code = NotFound desc = could not find container \"72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6\": container with ID starting with 72f9682f6ef2782f325e4bedbea6deaa9bafbfa4f7b03a547440cc35e2cceef6 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.345101 4876 scope.go:117] "RemoveContainer" containerID="3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.346386 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305"} err="failed to get container status \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": rpc error: code = NotFound desc = could not find container \"3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305\": container with ID starting with 3d77c6e37f4f12491dbf661c0f253f362e848eeeea904ac1360dfa7b51f1b305 not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.346418 4876 scope.go:117] "RemoveContainer" containerID="93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.346781 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff"} err="failed to get container status \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": rpc error: code = NotFound desc = could not find container \"93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff\": container with ID starting with 93055a804cf037583669700053e18eb772898f273e3004ae3a9b2389c4ac8eff not found: ID does not exist" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.346815 4876 scope.go:117] "RemoveContainer" containerID="1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf" Mar 13 07:51:29 crc kubenswrapper[4876]: I0313 07:51:29.347112 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf"} err="failed to get container status \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": rpc error: code = NotFound desc = could not find container \"1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf\": container with ID starting with 1e795cbce469f3de21697e3a0432e9d9c7c3dfa42c6b3b83a0bceaa92888cacf not found: ID does not exist" Mar 13 07:51:30 crc kubenswrapper[4876]: I0313 07:51:30.037586 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/2.log" Mar 13 07:51:30 crc kubenswrapper[4876]: I0313 07:51:30.039400 4876 generic.go:334] "Generic (PLEG): container finished" podID="f22530c5-04ed-4284-bbff-61bb844d8c40" containerID="2eb8694301917035047b884c466d2691134128fe326dbbb15f46c13d1b84721a" exitCode=0 Mar 13 07:51:30 crc kubenswrapper[4876]: I0313 07:51:30.039493 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerDied","Data":"2eb8694301917035047b884c466d2691134128fe326dbbb15f46c13d1b84721a"} Mar 13 07:51:30 crc kubenswrapper[4876]: I0313 07:51:30.039593 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"282528fb294e0619bc4f7cd249031ad61adc0ae706a417548a1f8bc0bd63e64c"} Mar 13 07:51:31 crc kubenswrapper[4876]: I0313 07:51:31.051687 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd" path="/var/lib/kubelet/pods/6a2d8ff1-99c6-44ed-bb4d-cd753c9126fd/volumes" Mar 13 07:51:31 crc kubenswrapper[4876]: I0313 07:51:31.056181 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"da43d474ff92744e4f11ed43407c01198eed8b4149a06314915b7f8295008efc"} Mar 13 07:51:31 crc kubenswrapper[4876]: I0313 07:51:31.056280 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"e67856f4a9bf7c08bb1defe1c926035a2bc5f8f67e871f7b855ae990c059b250"} Mar 13 07:51:31 crc kubenswrapper[4876]: I0313 07:51:31.056311 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"5543981c62b9a87327c66c3988ac01c70a2a79e3eba22ba2e0d0a4af9c8a796e"} Mar 13 07:51:31 crc kubenswrapper[4876]: I0313 07:51:31.056340 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"739bca1f5ac88ff21421f1faa12f79ac2cb25130d78c9bb6c25bdd032edc2004"} Mar 13 07:51:31 crc kubenswrapper[4876]: I0313 07:51:31.056365 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"d78ef3567385f30926253807425ddd2886372da7cf870b3175cdafa26a77afd9"} Mar 13 07:51:31 crc kubenswrapper[4876]: I0313 07:51:31.056389 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"d5aebbf9d19e4584da747be80464d86e71740aa70f1ce8c5a79776b04d327cbc"} Mar 13 07:51:34 crc kubenswrapper[4876]: I0313 07:51:34.094970 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"f24b6721ec7b9d73e29c8b991b5cc24a5569773b76d5e8321550ed5340705f69"} Mar 13 07:51:36 crc kubenswrapper[4876]: I0313 07:51:36.112004 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" event={"ID":"f22530c5-04ed-4284-bbff-61bb844d8c40","Type":"ContainerStarted","Data":"b64d92afc02e62733342889072640cd3b178d3c052dbec6f7f78925d0919470d"} Mar 13 07:51:36 crc kubenswrapper[4876]: I0313 07:51:36.112582 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:36 crc kubenswrapper[4876]: I0313 07:51:36.112599 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:36 crc kubenswrapper[4876]: I0313 07:51:36.112611 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:36 crc kubenswrapper[4876]: I0313 07:51:36.142663 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:36 crc kubenswrapper[4876]: I0313 07:51:36.145685 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:51:36 crc kubenswrapper[4876]: I0313 07:51:36.148221 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" podStartSLOduration=8.1482045 podStartE2EDuration="8.1482045s" podCreationTimestamp="2026-03-13 07:51:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:51:36.144336739 +0000 UTC m=+755.815115731" watchObservedRunningTime="2026-03-13 07:51:36.1482045 +0000 UTC m=+755.818983482" Mar 13 07:51:43 crc kubenswrapper[4876]: I0313 07:51:43.036308 4876 scope.go:117] "RemoveContainer" containerID="76512308b59931886eee75e5bd8c0ff26eed1bfe288f2ff41e80a8ca4761ad73" Mar 13 07:51:43 crc kubenswrapper[4876]: E0313 07:51:43.037418 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-st667_openshift-multus(8adb2a69-d0ef-4efc-813d-77fdf9713d50)\"" pod="openshift-multus/multus-st667" podUID="8adb2a69-d0ef-4efc-813d-77fdf9713d50" Mar 13 07:51:54 crc kubenswrapper[4876]: I0313 07:51:54.035578 4876 scope.go:117] "RemoveContainer" containerID="76512308b59931886eee75e5bd8c0ff26eed1bfe288f2ff41e80a8ca4761ad73" Mar 13 07:51:54 crc kubenswrapper[4876]: I0313 07:51:54.234120 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-st667_8adb2a69-d0ef-4efc-813d-77fdf9713d50/kube-multus/2.log" Mar 13 07:51:54 crc kubenswrapper[4876]: I0313 07:51:54.234750 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-st667" event={"ID":"8adb2a69-d0ef-4efc-813d-77fdf9713d50","Type":"ContainerStarted","Data":"95dd1e611b43c91870178ac11cb7b9413ee0556d798bd6ce0c113140eff12159"} Mar 13 07:51:59 crc kubenswrapper[4876]: I0313 07:51:59.107656 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9bxnt" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.152556 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556472-68q8k"] Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.153441 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556472-68q8k" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.160133 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.160557 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.160802 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.170740 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556472-68q8k"] Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.223089 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdd5p\" (UniqueName: \"kubernetes.io/projected/710597ba-1b67-45a8-83de-8c9a23cccd4c-kube-api-access-qdd5p\") pod \"auto-csr-approver-29556472-68q8k\" (UID: \"710597ba-1b67-45a8-83de-8c9a23cccd4c\") " pod="openshift-infra/auto-csr-approver-29556472-68q8k" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.325041 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdd5p\" (UniqueName: \"kubernetes.io/projected/710597ba-1b67-45a8-83de-8c9a23cccd4c-kube-api-access-qdd5p\") pod \"auto-csr-approver-29556472-68q8k\" (UID: \"710597ba-1b67-45a8-83de-8c9a23cccd4c\") " pod="openshift-infra/auto-csr-approver-29556472-68q8k" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.353712 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdd5p\" (UniqueName: \"kubernetes.io/projected/710597ba-1b67-45a8-83de-8c9a23cccd4c-kube-api-access-qdd5p\") pod \"auto-csr-approver-29556472-68q8k\" (UID: \"710597ba-1b67-45a8-83de-8c9a23cccd4c\") " pod="openshift-infra/auto-csr-approver-29556472-68q8k" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.475602 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556472-68q8k" Mar 13 07:52:00 crc kubenswrapper[4876]: I0313 07:52:00.954042 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556472-68q8k"] Mar 13 07:52:01 crc kubenswrapper[4876]: I0313 07:52:01.279586 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556472-68q8k" event={"ID":"710597ba-1b67-45a8-83de-8c9a23cccd4c","Type":"ContainerStarted","Data":"94af9ef34c93d4158fbda8d51cbf9a8e1cff28b9e0e17241faa1bdecaa1a9cbd"} Mar 13 07:52:03 crc kubenswrapper[4876]: I0313 07:52:03.296850 4876 generic.go:334] "Generic (PLEG): container finished" podID="710597ba-1b67-45a8-83de-8c9a23cccd4c" containerID="1bd84d48a8e6bbb46eff26ac0a20237dbee3e8d502a155a3800850ade3a54e24" exitCode=0 Mar 13 07:52:03 crc kubenswrapper[4876]: I0313 07:52:03.296925 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556472-68q8k" event={"ID":"710597ba-1b67-45a8-83de-8c9a23cccd4c","Type":"ContainerDied","Data":"1bd84d48a8e6bbb46eff26ac0a20237dbee3e8d502a155a3800850ade3a54e24"} Mar 13 07:52:04 crc kubenswrapper[4876]: I0313 07:52:04.603354 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556472-68q8k" Mar 13 07:52:04 crc kubenswrapper[4876]: I0313 07:52:04.785770 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdd5p\" (UniqueName: \"kubernetes.io/projected/710597ba-1b67-45a8-83de-8c9a23cccd4c-kube-api-access-qdd5p\") pod \"710597ba-1b67-45a8-83de-8c9a23cccd4c\" (UID: \"710597ba-1b67-45a8-83de-8c9a23cccd4c\") " Mar 13 07:52:04 crc kubenswrapper[4876]: I0313 07:52:04.801959 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/710597ba-1b67-45a8-83de-8c9a23cccd4c-kube-api-access-qdd5p" (OuterVolumeSpecName: "kube-api-access-qdd5p") pod "710597ba-1b67-45a8-83de-8c9a23cccd4c" (UID: "710597ba-1b67-45a8-83de-8c9a23cccd4c"). InnerVolumeSpecName "kube-api-access-qdd5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:52:04 crc kubenswrapper[4876]: I0313 07:52:04.888499 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdd5p\" (UniqueName: \"kubernetes.io/projected/710597ba-1b67-45a8-83de-8c9a23cccd4c-kube-api-access-qdd5p\") on node \"crc\" DevicePath \"\"" Mar 13 07:52:05 crc kubenswrapper[4876]: I0313 07:52:05.314119 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556472-68q8k" event={"ID":"710597ba-1b67-45a8-83de-8c9a23cccd4c","Type":"ContainerDied","Data":"94af9ef34c93d4158fbda8d51cbf9a8e1cff28b9e0e17241faa1bdecaa1a9cbd"} Mar 13 07:52:05 crc kubenswrapper[4876]: I0313 07:52:05.314191 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556472-68q8k" Mar 13 07:52:05 crc kubenswrapper[4876]: I0313 07:52:05.314205 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94af9ef34c93d4158fbda8d51cbf9a8e1cff28b9e0e17241faa1bdecaa1a9cbd" Mar 13 07:52:05 crc kubenswrapper[4876]: I0313 07:52:05.676967 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556466-zhrhb"] Mar 13 07:52:05 crc kubenswrapper[4876]: I0313 07:52:05.680365 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556466-zhrhb"] Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.486643 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt"] Mar 13 07:52:06 crc kubenswrapper[4876]: E0313 07:52:06.486886 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="710597ba-1b67-45a8-83de-8c9a23cccd4c" containerName="oc" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.486898 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="710597ba-1b67-45a8-83de-8c9a23cccd4c" containerName="oc" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.487034 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="710597ba-1b67-45a8-83de-8c9a23cccd4c" containerName="oc" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.487794 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.490636 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.499591 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt"] Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.615342 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.615413 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kfxt\" (UniqueName: \"kubernetes.io/projected/ea82483b-90c0-4edd-b51c-38d2ae56de9e-kube-api-access-4kfxt\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.616034 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.717838 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.717914 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kfxt\" (UniqueName: \"kubernetes.io/projected/ea82483b-90c0-4edd-b51c-38d2ae56de9e-kube-api-access-4kfxt\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.718031 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.718499 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.718632 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.740689 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kfxt\" (UniqueName: \"kubernetes.io/projected/ea82483b-90c0-4edd-b51c-38d2ae56de9e-kube-api-access-4kfxt\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:06 crc kubenswrapper[4876]: I0313 07:52:06.861665 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:07 crc kubenswrapper[4876]: I0313 07:52:07.051130 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ace24c1-074f-4c5b-8148-5bb46a790a3e" path="/var/lib/kubelet/pods/5ace24c1-074f-4c5b-8148-5bb46a790a3e/volumes" Mar 13 07:52:07 crc kubenswrapper[4876]: I0313 07:52:07.127143 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt"] Mar 13 07:52:07 crc kubenswrapper[4876]: W0313 07:52:07.135078 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea82483b_90c0_4edd_b51c_38d2ae56de9e.slice/crio-f1ff7d78a0200b46e01de3dc40e2eb611415117ac78239e24c6117af73cfdcca WatchSource:0}: Error finding container f1ff7d78a0200b46e01de3dc40e2eb611415117ac78239e24c6117af73cfdcca: Status 404 returned error can't find the container with id f1ff7d78a0200b46e01de3dc40e2eb611415117ac78239e24c6117af73cfdcca Mar 13 07:52:07 crc kubenswrapper[4876]: I0313 07:52:07.329977 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" event={"ID":"ea82483b-90c0-4edd-b51c-38d2ae56de9e","Type":"ContainerStarted","Data":"d4e60552847925f2d85f175fb4aa9c35e60ae0f6d9074c5b829dfee7f6d266c3"} Mar 13 07:52:07 crc kubenswrapper[4876]: I0313 07:52:07.330581 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" event={"ID":"ea82483b-90c0-4edd-b51c-38d2ae56de9e","Type":"ContainerStarted","Data":"f1ff7d78a0200b46e01de3dc40e2eb611415117ac78239e24c6117af73cfdcca"} Mar 13 07:52:08 crc kubenswrapper[4876]: I0313 07:52:08.338740 4876 generic.go:334] "Generic (PLEG): container finished" podID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerID="d4e60552847925f2d85f175fb4aa9c35e60ae0f6d9074c5b829dfee7f6d266c3" exitCode=0 Mar 13 07:52:08 crc kubenswrapper[4876]: I0313 07:52:08.338838 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" event={"ID":"ea82483b-90c0-4edd-b51c-38d2ae56de9e","Type":"ContainerDied","Data":"d4e60552847925f2d85f175fb4aa9c35e60ae0f6d9074c5b829dfee7f6d266c3"} Mar 13 07:52:10 crc kubenswrapper[4876]: I0313 07:52:10.358397 4876 generic.go:334] "Generic (PLEG): container finished" podID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerID="f8a1a37817518d65bace2bf47224d7c119a3ebe3487ff35bd0fa13754110abeb" exitCode=0 Mar 13 07:52:10 crc kubenswrapper[4876]: I0313 07:52:10.358488 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" event={"ID":"ea82483b-90c0-4edd-b51c-38d2ae56de9e","Type":"ContainerDied","Data":"f8a1a37817518d65bace2bf47224d7c119a3ebe3487ff35bd0fa13754110abeb"} Mar 13 07:52:11 crc kubenswrapper[4876]: I0313 07:52:11.368319 4876 generic.go:334] "Generic (PLEG): container finished" podID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerID="7325bb1c37680152c42d15b4b3aaf029b354b3485a0ecd3d56ea8fbbf89ce5db" exitCode=0 Mar 13 07:52:11 crc kubenswrapper[4876]: I0313 07:52:11.368765 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" event={"ID":"ea82483b-90c0-4edd-b51c-38d2ae56de9e","Type":"ContainerDied","Data":"7325bb1c37680152c42d15b4b3aaf029b354b3485a0ecd3d56ea8fbbf89ce5db"} Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.621187 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.810451 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-bundle\") pod \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.810687 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-util\") pod \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.810842 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kfxt\" (UniqueName: \"kubernetes.io/projected/ea82483b-90c0-4edd-b51c-38d2ae56de9e-kube-api-access-4kfxt\") pod \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\" (UID: \"ea82483b-90c0-4edd-b51c-38d2ae56de9e\") " Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.811696 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-bundle" (OuterVolumeSpecName: "bundle") pod "ea82483b-90c0-4edd-b51c-38d2ae56de9e" (UID: "ea82483b-90c0-4edd-b51c-38d2ae56de9e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.820937 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea82483b-90c0-4edd-b51c-38d2ae56de9e-kube-api-access-4kfxt" (OuterVolumeSpecName: "kube-api-access-4kfxt") pod "ea82483b-90c0-4edd-b51c-38d2ae56de9e" (UID: "ea82483b-90c0-4edd-b51c-38d2ae56de9e"). InnerVolumeSpecName "kube-api-access-4kfxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.822738 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-util" (OuterVolumeSpecName: "util") pod "ea82483b-90c0-4edd-b51c-38d2ae56de9e" (UID: "ea82483b-90c0-4edd-b51c-38d2ae56de9e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.912186 4876 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-util\") on node \"crc\" DevicePath \"\"" Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.912254 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kfxt\" (UniqueName: \"kubernetes.io/projected/ea82483b-90c0-4edd-b51c-38d2ae56de9e-kube-api-access-4kfxt\") on node \"crc\" DevicePath \"\"" Mar 13 07:52:12 crc kubenswrapper[4876]: I0313 07:52:12.912274 4876 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea82483b-90c0-4edd-b51c-38d2ae56de9e-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:52:13 crc kubenswrapper[4876]: I0313 07:52:13.386092 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" event={"ID":"ea82483b-90c0-4edd-b51c-38d2ae56de9e","Type":"ContainerDied","Data":"f1ff7d78a0200b46e01de3dc40e2eb611415117ac78239e24c6117af73cfdcca"} Mar 13 07:52:13 crc kubenswrapper[4876]: I0313 07:52:13.386190 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1ff7d78a0200b46e01de3dc40e2eb611415117ac78239e24c6117af73cfdcca" Mar 13 07:52:13 crc kubenswrapper[4876]: I0313 07:52:13.386311 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.882599 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s"] Mar 13 07:52:17 crc kubenswrapper[4876]: E0313 07:52:17.883198 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerName="util" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.883216 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerName="util" Mar 13 07:52:17 crc kubenswrapper[4876]: E0313 07:52:17.883228 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerName="pull" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.883234 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerName="pull" Mar 13 07:52:17 crc kubenswrapper[4876]: E0313 07:52:17.883261 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerName="extract" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.883268 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerName="extract" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.883377 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea82483b-90c0-4edd-b51c-38d2ae56de9e" containerName="extract" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.883782 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.887646 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.887837 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bkvs6" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.897229 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.901449 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s"] Mar 13 07:52:17 crc kubenswrapper[4876]: I0313 07:52:17.986342 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txjjm\" (UniqueName: \"kubernetes.io/projected/54e0cdfc-3234-4763-bc49-6a6e7a537c74-kube-api-access-txjjm\") pod \"nmstate-operator-796d4cfff4-5gv4s\" (UID: \"54e0cdfc-3234-4763-bc49-6a6e7a537c74\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s" Mar 13 07:52:18 crc kubenswrapper[4876]: I0313 07:52:18.088074 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txjjm\" (UniqueName: \"kubernetes.io/projected/54e0cdfc-3234-4763-bc49-6a6e7a537c74-kube-api-access-txjjm\") pod \"nmstate-operator-796d4cfff4-5gv4s\" (UID: \"54e0cdfc-3234-4763-bc49-6a6e7a537c74\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s" Mar 13 07:52:18 crc kubenswrapper[4876]: I0313 07:52:18.107699 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txjjm\" (UniqueName: \"kubernetes.io/projected/54e0cdfc-3234-4763-bc49-6a6e7a537c74-kube-api-access-txjjm\") pod \"nmstate-operator-796d4cfff4-5gv4s\" (UID: \"54e0cdfc-3234-4763-bc49-6a6e7a537c74\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s" Mar 13 07:52:18 crc kubenswrapper[4876]: I0313 07:52:18.200828 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s" Mar 13 07:52:18 crc kubenswrapper[4876]: I0313 07:52:18.454494 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s"] Mar 13 07:52:19 crc kubenswrapper[4876]: I0313 07:52:19.426034 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s" event={"ID":"54e0cdfc-3234-4763-bc49-6a6e7a537c74","Type":"ContainerStarted","Data":"4eedf5110751f30710284cb029eb1993b2f58aea2ab8ff148dc3f28f9f57575a"} Mar 13 07:52:22 crc kubenswrapper[4876]: I0313 07:52:22.451925 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s" event={"ID":"54e0cdfc-3234-4763-bc49-6a6e7a537c74","Type":"ContainerStarted","Data":"9730a2058602bbc2f9a0cfe5883b60ee7b1fab74f7941730e5e8b143b5be0393"} Mar 13 07:52:22 crc kubenswrapper[4876]: I0313 07:52:22.476003 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-5gv4s" podStartSLOduration=2.449886308 podStartE2EDuration="5.475972718s" podCreationTimestamp="2026-03-13 07:52:17 +0000 UTC" firstStartedPulling="2026-03-13 07:52:18.470192711 +0000 UTC m=+798.140971693" lastFinishedPulling="2026-03-13 07:52:21.496279121 +0000 UTC m=+801.167058103" observedRunningTime="2026-03-13 07:52:22.472341944 +0000 UTC m=+802.143120956" watchObservedRunningTime="2026-03-13 07:52:22.475972718 +0000 UTC m=+802.146751740" Mar 13 07:52:24 crc kubenswrapper[4876]: I0313 07:52:24.608006 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:52:24 crc kubenswrapper[4876]: I0313 07:52:24.608777 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.255364 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg"] Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.257715 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.260996 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-74bqr"] Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.262452 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.263651 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-pkvdx" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.269865 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.274353 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-74bqr"] Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.278509 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg"] Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.290987 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-v9cql"] Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.292159 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.302669 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.302727 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-ovs-socket\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.302749 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h2kn\" (UniqueName: \"kubernetes.io/projected/8f70b2a8-97e0-476b-91d2-e5c386d8a14f-kube-api-access-9h2kn\") pod \"nmstate-metrics-9b8c8685d-d2mlg\" (UID: \"8f70b2a8-97e0-476b-91d2-e5c386d8a14f\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.302764 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh787\" (UniqueName: \"kubernetes.io/projected/e506e0e1-25c2-4be3-ad95-5f9029a960e3-kube-api-access-hh787\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.302787 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-nmstate-lock\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.302809 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcmwj\" (UniqueName: \"kubernetes.io/projected/6aa5c930-3567-401a-830f-c30ca123b57a-kube-api-access-mcmwj\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.302848 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-dbus-socket\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.403786 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h2kn\" (UniqueName: \"kubernetes.io/projected/8f70b2a8-97e0-476b-91d2-e5c386d8a14f-kube-api-access-9h2kn\") pod \"nmstate-metrics-9b8c8685d-d2mlg\" (UID: \"8f70b2a8-97e0-476b-91d2-e5c386d8a14f\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.403839 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh787\" (UniqueName: \"kubernetes.io/projected/e506e0e1-25c2-4be3-ad95-5f9029a960e3-kube-api-access-hh787\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.403875 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-nmstate-lock\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.403904 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcmwj\" (UniqueName: \"kubernetes.io/projected/6aa5c930-3567-401a-830f-c30ca123b57a-kube-api-access-mcmwj\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.403951 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-dbus-socket\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.403991 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.404035 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-ovs-socket\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.404098 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-ovs-socket\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.404102 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-nmstate-lock\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: E0313 07:52:32.404291 4876 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Mar 13 07:52:32 crc kubenswrapper[4876]: E0313 07:52:32.404378 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair podName:e506e0e1-25c2-4be3-ad95-5f9029a960e3 nodeName:}" failed. No retries permitted until 2026-03-13 07:52:32.904349264 +0000 UTC m=+812.575128246 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair") pod "nmstate-webhook-5f558f5558-74bqr" (UID: "e506e0e1-25c2-4be3-ad95-5f9029a960e3") : secret "openshift-nmstate-webhook" not found Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.404574 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6aa5c930-3567-401a-830f-c30ca123b57a-dbus-socket\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.420798 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj"] Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.421610 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.427174 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7m22l" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.427445 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.427971 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.428520 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcmwj\" (UniqueName: \"kubernetes.io/projected/6aa5c930-3567-401a-830f-c30ca123b57a-kube-api-access-mcmwj\") pod \"nmstate-handler-v9cql\" (UID: \"6aa5c930-3567-401a-830f-c30ca123b57a\") " pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.428901 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h2kn\" (UniqueName: \"kubernetes.io/projected/8f70b2a8-97e0-476b-91d2-e5c386d8a14f-kube-api-access-9h2kn\") pod \"nmstate-metrics-9b8c8685d-d2mlg\" (UID: \"8f70b2a8-97e0-476b-91d2-e5c386d8a14f\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.432246 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj"] Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.436358 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh787\" (UniqueName: \"kubernetes.io/projected/e506e0e1-25c2-4be3-ad95-5f9029a960e3-kube-api-access-hh787\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.504899 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/be89d0d3-8483-4439-b346-c50bd885b15c-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.505507 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p7zb\" (UniqueName: \"kubernetes.io/projected/be89d0d3-8483-4439-b346-c50bd885b15c-kube-api-access-2p7zb\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.505543 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.578646 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.606844 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p7zb\" (UniqueName: \"kubernetes.io/projected/be89d0d3-8483-4439-b346-c50bd885b15c-kube-api-access-2p7zb\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.606951 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.607011 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/be89d0d3-8483-4439-b346-c50bd885b15c-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:32 crc kubenswrapper[4876]: E0313 07:52:32.607180 4876 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 13 07:52:32 crc kubenswrapper[4876]: E0313 07:52:32.607290 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert podName:be89d0d3-8483-4439-b346-c50bd885b15c nodeName:}" failed. No retries permitted until 2026-03-13 07:52:33.107265529 +0000 UTC m=+812.778044511 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-62blj" (UID: "be89d0d3-8483-4439-b346-c50bd885b15c") : secret "plugin-serving-cert" not found Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.608638 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/be89d0d3-8483-4439-b346-c50bd885b15c-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.611578 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:32 crc kubenswrapper[4876]: I0313 07:52:32.909985 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:32 crc kubenswrapper[4876]: E0313 07:52:32.910172 4876 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Mar 13 07:52:32 crc kubenswrapper[4876]: E0313 07:52:32.910309 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair podName:e506e0e1-25c2-4be3-ad95-5f9029a960e3 nodeName:}" failed. No retries permitted until 2026-03-13 07:52:33.910287466 +0000 UTC m=+813.581066448 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair") pod "nmstate-webhook-5f558f5558-74bqr" (UID: "e506e0e1-25c2-4be3-ad95-5f9029a960e3") : secret "openshift-nmstate-webhook" not found Mar 13 07:52:33 crc kubenswrapper[4876]: I0313 07:52:33.112419 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:33 crc kubenswrapper[4876]: E0313 07:52:33.112644 4876 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 13 07:52:33 crc kubenswrapper[4876]: E0313 07:52:33.112726 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert podName:be89d0d3-8483-4439-b346-c50bd885b15c nodeName:}" failed. No retries permitted until 2026-03-13 07:52:34.112704156 +0000 UTC m=+813.783483138 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-62blj" (UID: "be89d0d3-8483-4439-b346-c50bd885b15c") : secret "plugin-serving-cert" not found Mar 13 07:52:33 crc kubenswrapper[4876]: I0313 07:52:33.534849 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-v9cql" event={"ID":"6aa5c930-3567-401a-830f-c30ca123b57a","Type":"ContainerStarted","Data":"fd00eb38e3d5bee3fe204d1dbbef8d7e2bc502331701b802ea76956725cd8269"} Mar 13 07:52:33 crc kubenswrapper[4876]: I0313 07:52:33.717801 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-zvfb8" podUID="729a14e7-a89b-4e7c-a8b5-a4002519d6a3" containerName="registry-server" probeResult="failure" output=< Mar 13 07:52:33 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 07:52:33 crc kubenswrapper[4876]: > Mar 13 07:52:33 crc kubenswrapper[4876]: I0313 07:52:33.752756 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-zvfb8" podUID="729a14e7-a89b-4e7c-a8b5-a4002519d6a3" containerName="registry-server" probeResult="failure" output=< Mar 13 07:52:33 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 07:52:33 crc kubenswrapper[4876]: > Mar 13 07:52:33 crc kubenswrapper[4876]: I0313 07:52:33.927104 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:33 crc kubenswrapper[4876]: E0313 07:52:33.927318 4876 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Mar 13 07:52:33 crc kubenswrapper[4876]: E0313 07:52:33.927444 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair podName:e506e0e1-25c2-4be3-ad95-5f9029a960e3 nodeName:}" failed. No retries permitted until 2026-03-13 07:52:35.927423031 +0000 UTC m=+815.598202013 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair") pod "nmstate-webhook-5f558f5558-74bqr" (UID: "e506e0e1-25c2-4be3-ad95-5f9029a960e3") : secret "openshift-nmstate-webhook" not found Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.008121 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p7zb\" (UniqueName: \"kubernetes.io/projected/be89d0d3-8483-4439-b346-c50bd885b15c-kube-api-access-2p7zb\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.127736 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-779c478558-qz6dr"] Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.131497 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.140103 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-config\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.140161 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgtlm\" (UniqueName: \"kubernetes.io/projected/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-kube-api-access-xgtlm\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.140221 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-trusted-ca-bundle\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.140286 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-oauth-config\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.140339 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-service-ca\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.140374 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.140397 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-serving-cert\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.140425 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-oauth-serving-cert\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.143111 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-779c478558-qz6dr"] Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.147926 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be89d0d3-8483-4439-b346-c50bd885b15c-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-62blj\" (UID: \"be89d0d3-8483-4439-b346-c50bd885b15c\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.242978 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgtlm\" (UniqueName: \"kubernetes.io/projected/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-kube-api-access-xgtlm\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.243692 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-config\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.243759 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-trusted-ca-bundle\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.243817 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-oauth-config\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.243882 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-service-ca\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.243917 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-serving-cert\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.243946 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-oauth-serving-cert\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.245023 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-config\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.246553 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-service-ca\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.247750 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-serving-cert\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.247850 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-console-oauth-config\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.248259 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-trusted-ca-bundle\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.249130 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-oauth-serving-cert\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.264816 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgtlm\" (UniqueName: \"kubernetes.io/projected/b53a8f92-32a5-4d6e-b9b0-e15e8847bce3-kube-api-access-xgtlm\") pod \"console-779c478558-qz6dr\" (UID: \"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3\") " pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.275562 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.423831 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg"] Mar 13 07:52:34 crc kubenswrapper[4876]: W0313 07:52:34.435266 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f70b2a8_97e0_476b_91d2_e5c386d8a14f.slice/crio-b045dc1fb8260e3d26f50f3b598c3cd60f168899305cce1ab9206eafe3035ab9 WatchSource:0}: Error finding container b045dc1fb8260e3d26f50f3b598c3cd60f168899305cce1ab9206eafe3035ab9: Status 404 returned error can't find the container with id b045dc1fb8260e3d26f50f3b598c3cd60f168899305cce1ab9206eafe3035ab9 Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.477947 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.499948 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj"] Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.546524 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" event={"ID":"be89d0d3-8483-4439-b346-c50bd885b15c","Type":"ContainerStarted","Data":"37fa70e221ab3bc777b47da96162963df67bf4813bce5a2223ab05c4b5af5901"} Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.547306 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" event={"ID":"8f70b2a8-97e0-476b-91d2-e5c386d8a14f","Type":"ContainerStarted","Data":"b045dc1fb8260e3d26f50f3b598c3cd60f168899305cce1ab9206eafe3035ab9"} Mar 13 07:52:34 crc kubenswrapper[4876]: I0313 07:52:34.691162 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-779c478558-qz6dr"] Mar 13 07:52:34 crc kubenswrapper[4876]: W0313 07:52:34.702631 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb53a8f92_32a5_4d6e_b9b0_e15e8847bce3.slice/crio-aaec738af76f6d3dc328d4fbc6cffeabe5cebb9df193146bb701c6e88cc27d10 WatchSource:0}: Error finding container aaec738af76f6d3dc328d4fbc6cffeabe5cebb9df193146bb701c6e88cc27d10: Status 404 returned error can't find the container with id aaec738af76f6d3dc328d4fbc6cffeabe5cebb9df193146bb701c6e88cc27d10 Mar 13 07:52:35 crc kubenswrapper[4876]: I0313 07:52:35.557054 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-779c478558-qz6dr" event={"ID":"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3","Type":"ContainerStarted","Data":"582886eb7a26d3aad53842f4ec95614d6e93bd9b148df33272b83011f14dcd58"} Mar 13 07:52:35 crc kubenswrapper[4876]: I0313 07:52:35.560491 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-779c478558-qz6dr" event={"ID":"b53a8f92-32a5-4d6e-b9b0-e15e8847bce3","Type":"ContainerStarted","Data":"aaec738af76f6d3dc328d4fbc6cffeabe5cebb9df193146bb701c6e88cc27d10"} Mar 13 07:52:35 crc kubenswrapper[4876]: I0313 07:52:35.581858 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-779c478558-qz6dr" podStartSLOduration=1.58182174 podStartE2EDuration="1.58182174s" podCreationTimestamp="2026-03-13 07:52:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:52:35.5744457 +0000 UTC m=+815.245224682" watchObservedRunningTime="2026-03-13 07:52:35.58182174 +0000 UTC m=+815.252600762" Mar 13 07:52:35 crc kubenswrapper[4876]: I0313 07:52:35.968518 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:35 crc kubenswrapper[4876]: I0313 07:52:35.980199 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e506e0e1-25c2-4be3-ad95-5f9029a960e3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-74bqr\" (UID: \"e506e0e1-25c2-4be3-ad95-5f9029a960e3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:36 crc kubenswrapper[4876]: I0313 07:52:36.186844 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:36 crc kubenswrapper[4876]: I0313 07:52:36.565792 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-v9cql" event={"ID":"6aa5c930-3567-401a-830f-c30ca123b57a","Type":"ContainerStarted","Data":"102c6768f808563797eac957838d7752d1cb9189ea62640ca25f7507ee05f5d1"} Mar 13 07:52:36 crc kubenswrapper[4876]: I0313 07:52:36.565993 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:36 crc kubenswrapper[4876]: I0313 07:52:36.569961 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" event={"ID":"8f70b2a8-97e0-476b-91d2-e5c386d8a14f","Type":"ContainerStarted","Data":"919a287cbea75d5dcab7472460cbdff5c04e1de64e4d5bb022e8eccc438f5149"} Mar 13 07:52:36 crc kubenswrapper[4876]: I0313 07:52:36.896429 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-v9cql" podStartSLOduration=1.9024689879999999 podStartE2EDuration="4.896407503s" podCreationTimestamp="2026-03-13 07:52:32 +0000 UTC" firstStartedPulling="2026-03-13 07:52:32.632149387 +0000 UTC m=+812.302928369" lastFinishedPulling="2026-03-13 07:52:35.626087902 +0000 UTC m=+815.296866884" observedRunningTime="2026-03-13 07:52:36.588533447 +0000 UTC m=+816.259312439" watchObservedRunningTime="2026-03-13 07:52:36.896407503 +0000 UTC m=+816.567186505" Mar 13 07:52:36 crc kubenswrapper[4876]: I0313 07:52:36.899389 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-74bqr"] Mar 13 07:52:36 crc kubenswrapper[4876]: W0313 07:52:36.910018 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode506e0e1_25c2_4be3_ad95_5f9029a960e3.slice/crio-ceb696f9c930b3fdabd099629f88d2adc13b43e7f5f419128c972ba62b3f8473 WatchSource:0}: Error finding container ceb696f9c930b3fdabd099629f88d2adc13b43e7f5f419128c972ba62b3f8473: Status 404 returned error can't find the container with id ceb696f9c930b3fdabd099629f88d2adc13b43e7f5f419128c972ba62b3f8473 Mar 13 07:52:37 crc kubenswrapper[4876]: I0313 07:52:37.580406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" event={"ID":"be89d0d3-8483-4439-b346-c50bd885b15c","Type":"ContainerStarted","Data":"7ac70d215eb54e5b501e2b5c4397b2028b0a7f3af4b063cac1be607cc90ac4e2"} Mar 13 07:52:37 crc kubenswrapper[4876]: I0313 07:52:37.583214 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" event={"ID":"e506e0e1-25c2-4be3-ad95-5f9029a960e3","Type":"ContainerStarted","Data":"ceb696f9c930b3fdabd099629f88d2adc13b43e7f5f419128c972ba62b3f8473"} Mar 13 07:52:37 crc kubenswrapper[4876]: I0313 07:52:37.600570 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-62blj" podStartSLOduration=3.363474346 podStartE2EDuration="5.600545645s" podCreationTimestamp="2026-03-13 07:52:32 +0000 UTC" firstStartedPulling="2026-03-13 07:52:34.510556943 +0000 UTC m=+814.181335915" lastFinishedPulling="2026-03-13 07:52:36.747628232 +0000 UTC m=+816.418407214" observedRunningTime="2026-03-13 07:52:37.600177815 +0000 UTC m=+817.270956797" watchObservedRunningTime="2026-03-13 07:52:37.600545645 +0000 UTC m=+817.271324627" Mar 13 07:52:38 crc kubenswrapper[4876]: I0313 07:52:38.592013 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" event={"ID":"8f70b2a8-97e0-476b-91d2-e5c386d8a14f","Type":"ContainerStarted","Data":"06c4a217fe42efb378b8df5fdf8d1a1fb66e2eb75c4c9399a4c8ba08af1d8b32"} Mar 13 07:52:38 crc kubenswrapper[4876]: I0313 07:52:38.596203 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" event={"ID":"e506e0e1-25c2-4be3-ad95-5f9029a960e3","Type":"ContainerStarted","Data":"4f511eda31d5c47d2601f350b6436ea04cb983f27a40d2296263f2361417a509"} Mar 13 07:52:38 crc kubenswrapper[4876]: I0313 07:52:38.617638 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-d2mlg" podStartSLOduration=2.885774019 podStartE2EDuration="6.617610487s" podCreationTimestamp="2026-03-13 07:52:32 +0000 UTC" firstStartedPulling="2026-03-13 07:52:34.444631974 +0000 UTC m=+814.115410956" lastFinishedPulling="2026-03-13 07:52:38.176468442 +0000 UTC m=+817.847247424" observedRunningTime="2026-03-13 07:52:38.616400872 +0000 UTC m=+818.287179864" watchObservedRunningTime="2026-03-13 07:52:38.617610487 +0000 UTC m=+818.288389489" Mar 13 07:52:38 crc kubenswrapper[4876]: I0313 07:52:38.648057 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" podStartSLOduration=5.38708547 podStartE2EDuration="6.648036405s" podCreationTimestamp="2026-03-13 07:52:32 +0000 UTC" firstStartedPulling="2026-03-13 07:52:36.914468788 +0000 UTC m=+816.585247770" lastFinishedPulling="2026-03-13 07:52:38.175419723 +0000 UTC m=+817.846198705" observedRunningTime="2026-03-13 07:52:38.647415027 +0000 UTC m=+818.318194009" watchObservedRunningTime="2026-03-13 07:52:38.648036405 +0000 UTC m=+818.318815387" Mar 13 07:52:39 crc kubenswrapper[4876]: I0313 07:52:39.603715 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:52:42 crc kubenswrapper[4876]: I0313 07:52:42.640648 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-v9cql" Mar 13 07:52:44 crc kubenswrapper[4876]: I0313 07:52:44.478681 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:44 crc kubenswrapper[4876]: I0313 07:52:44.479541 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:44 crc kubenswrapper[4876]: I0313 07:52:44.487178 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:44 crc kubenswrapper[4876]: I0313 07:52:44.649727 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-779c478558-qz6dr" Mar 13 07:52:44 crc kubenswrapper[4876]: I0313 07:52:44.709393 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-g65sf"] Mar 13 07:52:54 crc kubenswrapper[4876]: I0313 07:52:54.607343 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:52:54 crc kubenswrapper[4876]: I0313 07:52:54.608306 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:52:56 crc kubenswrapper[4876]: I0313 07:52:56.196595 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-74bqr" Mar 13 07:53:00 crc kubenswrapper[4876]: I0313 07:53:00.201542 4876 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 07:53:01 crc kubenswrapper[4876]: I0313 07:53:01.641748 4876 scope.go:117] "RemoveContainer" containerID="670a49da83eeaa094431c2a1bd7be454953fd61b819d601b25a86ec9ed070bfb" Mar 13 07:53:09 crc kubenswrapper[4876]: I0313 07:53:09.764686 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-g65sf" podUID="62fdbab2-2be8-42cc-a936-a87ffd085a0e" containerName="console" containerID="cri-o://02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27" gracePeriod=15 Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.188642 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-g65sf_62fdbab2-2be8-42cc-a936-a87ffd085a0e/console/0.log" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.189382 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.320118 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-oauth-serving-cert\") pod \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.320209 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-serving-cert\") pod \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.320276 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-service-ca\") pod \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.320332 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c59f6\" (UniqueName: \"kubernetes.io/projected/62fdbab2-2be8-42cc-a936-a87ffd085a0e-kube-api-access-c59f6\") pod \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.320369 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-config\") pod \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.320404 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-oauth-config\") pod \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.320434 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-trusted-ca-bundle\") pod \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\" (UID: \"62fdbab2-2be8-42cc-a936-a87ffd085a0e\") " Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.321375 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "62fdbab2-2be8-42cc-a936-a87ffd085a0e" (UID: "62fdbab2-2be8-42cc-a936-a87ffd085a0e"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.321405 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "62fdbab2-2be8-42cc-a936-a87ffd085a0e" (UID: "62fdbab2-2be8-42cc-a936-a87ffd085a0e"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.321425 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-config" (OuterVolumeSpecName: "console-config") pod "62fdbab2-2be8-42cc-a936-a87ffd085a0e" (UID: "62fdbab2-2be8-42cc-a936-a87ffd085a0e"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.321488 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-service-ca" (OuterVolumeSpecName: "service-ca") pod "62fdbab2-2be8-42cc-a936-a87ffd085a0e" (UID: "62fdbab2-2be8-42cc-a936-a87ffd085a0e"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.328901 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "62fdbab2-2be8-42cc-a936-a87ffd085a0e" (UID: "62fdbab2-2be8-42cc-a936-a87ffd085a0e"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.329177 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62fdbab2-2be8-42cc-a936-a87ffd085a0e-kube-api-access-c59f6" (OuterVolumeSpecName: "kube-api-access-c59f6") pod "62fdbab2-2be8-42cc-a936-a87ffd085a0e" (UID: "62fdbab2-2be8-42cc-a936-a87ffd085a0e"). InnerVolumeSpecName "kube-api-access-c59f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.330624 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "62fdbab2-2be8-42cc-a936-a87ffd085a0e" (UID: "62fdbab2-2be8-42cc-a936-a87ffd085a0e"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.421970 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c59f6\" (UniqueName: \"kubernetes.io/projected/62fdbab2-2be8-42cc-a936-a87ffd085a0e-kube-api-access-c59f6\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.422023 4876 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.422036 4876 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.422049 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.422061 4876 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.422075 4876 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/62fdbab2-2be8-42cc-a936-a87ffd085a0e-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.422088 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/62fdbab2-2be8-42cc-a936-a87ffd085a0e-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.646869 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c"] Mar 13 07:53:10 crc kubenswrapper[4876]: E0313 07:53:10.647407 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fdbab2-2be8-42cc-a936-a87ffd085a0e" containerName="console" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.647447 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fdbab2-2be8-42cc-a936-a87ffd085a0e" containerName="console" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.647698 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fdbab2-2be8-42cc-a936-a87ffd085a0e" containerName="console" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.649886 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.652918 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.679844 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c"] Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.828354 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdknz\" (UniqueName: \"kubernetes.io/projected/489dadce-846b-4338-b396-1c80e256b2f8-kube-api-access-hdknz\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.828432 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.828668 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.886414 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-g65sf_62fdbab2-2be8-42cc-a936-a87ffd085a0e/console/0.log" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.886512 4876 generic.go:334] "Generic (PLEG): container finished" podID="62fdbab2-2be8-42cc-a936-a87ffd085a0e" containerID="02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27" exitCode=2 Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.886570 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g65sf" event={"ID":"62fdbab2-2be8-42cc-a936-a87ffd085a0e","Type":"ContainerDied","Data":"02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27"} Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.886621 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g65sf" event={"ID":"62fdbab2-2be8-42cc-a936-a87ffd085a0e","Type":"ContainerDied","Data":"7f68a1abd786e08ed55e9e6f82c3c8fb64fbb9e19e00d648e05cf0ab70d47a54"} Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.886658 4876 scope.go:117] "RemoveContainer" containerID="02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.886908 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g65sf" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.910992 4876 scope.go:117] "RemoveContainer" containerID="02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27" Mar 13 07:53:10 crc kubenswrapper[4876]: E0313 07:53:10.911915 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27\": container with ID starting with 02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27 not found: ID does not exist" containerID="02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.912058 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27"} err="failed to get container status \"02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27\": rpc error: code = NotFound desc = could not find container \"02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27\": container with ID starting with 02fe812adcc3b6989c79efa95d0cb9123f78fbc3b1514e26d51ae2e00445ef27 not found: ID does not exist" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.930296 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-g65sf"] Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.930450 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdknz\" (UniqueName: \"kubernetes.io/projected/489dadce-846b-4338-b396-1c80e256b2f8-kube-api-access-hdknz\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.930538 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.930626 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.931433 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.931755 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.934893 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-g65sf"] Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.954752 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdknz\" (UniqueName: \"kubernetes.io/projected/489dadce-846b-4338-b396-1c80e256b2f8-kube-api-access-hdknz\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:10 crc kubenswrapper[4876]: I0313 07:53:10.971654 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:11 crc kubenswrapper[4876]: I0313 07:53:11.049799 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62fdbab2-2be8-42cc-a936-a87ffd085a0e" path="/var/lib/kubelet/pods/62fdbab2-2be8-42cc-a936-a87ffd085a0e/volumes" Mar 13 07:53:11 crc kubenswrapper[4876]: I0313 07:53:11.509614 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c"] Mar 13 07:53:11 crc kubenswrapper[4876]: W0313 07:53:11.525375 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod489dadce_846b_4338_b396_1c80e256b2f8.slice/crio-724f7a632e0d857d77b337e4c1305948096297e9939537c25c5ff30165f49965 WatchSource:0}: Error finding container 724f7a632e0d857d77b337e4c1305948096297e9939537c25c5ff30165f49965: Status 404 returned error can't find the container with id 724f7a632e0d857d77b337e4c1305948096297e9939537c25c5ff30165f49965 Mar 13 07:53:11 crc kubenswrapper[4876]: I0313 07:53:11.901366 4876 generic.go:334] "Generic (PLEG): container finished" podID="489dadce-846b-4338-b396-1c80e256b2f8" containerID="dfa557f5ef10f584c02fc9f057358e59b106be3bbd13fd422f82daadd214f4aa" exitCode=0 Mar 13 07:53:11 crc kubenswrapper[4876]: I0313 07:53:11.901477 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" event={"ID":"489dadce-846b-4338-b396-1c80e256b2f8","Type":"ContainerDied","Data":"dfa557f5ef10f584c02fc9f057358e59b106be3bbd13fd422f82daadd214f4aa"} Mar 13 07:53:11 crc kubenswrapper[4876]: I0313 07:53:11.902037 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" event={"ID":"489dadce-846b-4338-b396-1c80e256b2f8","Type":"ContainerStarted","Data":"724f7a632e0d857d77b337e4c1305948096297e9939537c25c5ff30165f49965"} Mar 13 07:53:11 crc kubenswrapper[4876]: I0313 07:53:11.904838 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 07:53:12 crc kubenswrapper[4876]: I0313 07:53:12.990138 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pfrvt"] Mar 13 07:53:12 crc kubenswrapper[4876]: I0313 07:53:12.992019 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:12 crc kubenswrapper[4876]: I0313 07:53:12.998348 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pfrvt"] Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.067702 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbpt2\" (UniqueName: \"kubernetes.io/projected/29ad6925-f568-4472-a5b9-a2e2e1a135b8-kube-api-access-tbpt2\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.067809 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-utilities\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.067860 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-catalog-content\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.169261 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-utilities\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.169351 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-catalog-content\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.169427 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbpt2\" (UniqueName: \"kubernetes.io/projected/29ad6925-f568-4472-a5b9-a2e2e1a135b8-kube-api-access-tbpt2\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.169904 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-utilities\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.169968 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-catalog-content\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.202391 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbpt2\" (UniqueName: \"kubernetes.io/projected/29ad6925-f568-4472-a5b9-a2e2e1a135b8-kube-api-access-tbpt2\") pod \"redhat-operators-pfrvt\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.326621 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.639327 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pfrvt"] Mar 13 07:53:13 crc kubenswrapper[4876]: W0313 07:53:13.660358 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29ad6925_f568_4472_a5b9_a2e2e1a135b8.slice/crio-bf63f34186ad1266b9eafe5fe7d122a9d59fc811f3b2b13799372d641266ff9a WatchSource:0}: Error finding container bf63f34186ad1266b9eafe5fe7d122a9d59fc811f3b2b13799372d641266ff9a: Status 404 returned error can't find the container with id bf63f34186ad1266b9eafe5fe7d122a9d59fc811f3b2b13799372d641266ff9a Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.920001 4876 generic.go:334] "Generic (PLEG): container finished" podID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerID="932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686" exitCode=0 Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.920121 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfrvt" event={"ID":"29ad6925-f568-4472-a5b9-a2e2e1a135b8","Type":"ContainerDied","Data":"932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686"} Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.920185 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfrvt" event={"ID":"29ad6925-f568-4472-a5b9-a2e2e1a135b8","Type":"ContainerStarted","Data":"bf63f34186ad1266b9eafe5fe7d122a9d59fc811f3b2b13799372d641266ff9a"} Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.923865 4876 generic.go:334] "Generic (PLEG): container finished" podID="489dadce-846b-4338-b396-1c80e256b2f8" containerID="626a342a0bab01148c8596ffa5a01659af0bb99c02d96df22ee551e39505d411" exitCode=0 Mar 13 07:53:13 crc kubenswrapper[4876]: I0313 07:53:13.923917 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" event={"ID":"489dadce-846b-4338-b396-1c80e256b2f8","Type":"ContainerDied","Data":"626a342a0bab01148c8596ffa5a01659af0bb99c02d96df22ee551e39505d411"} Mar 13 07:53:14 crc kubenswrapper[4876]: I0313 07:53:14.934682 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfrvt" event={"ID":"29ad6925-f568-4472-a5b9-a2e2e1a135b8","Type":"ContainerStarted","Data":"be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f"} Mar 13 07:53:14 crc kubenswrapper[4876]: I0313 07:53:14.937871 4876 generic.go:334] "Generic (PLEG): container finished" podID="489dadce-846b-4338-b396-1c80e256b2f8" containerID="4d22f7919356ecb1c11003d7bf5804191c5168969221b369fc9809d27b97046f" exitCode=0 Mar 13 07:53:14 crc kubenswrapper[4876]: I0313 07:53:14.937915 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" event={"ID":"489dadce-846b-4338-b396-1c80e256b2f8","Type":"ContainerDied","Data":"4d22f7919356ecb1c11003d7bf5804191c5168969221b369fc9809d27b97046f"} Mar 13 07:53:15 crc kubenswrapper[4876]: I0313 07:53:15.950954 4876 generic.go:334] "Generic (PLEG): container finished" podID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerID="be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f" exitCode=0 Mar 13 07:53:15 crc kubenswrapper[4876]: I0313 07:53:15.951056 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfrvt" event={"ID":"29ad6925-f568-4472-a5b9-a2e2e1a135b8","Type":"ContainerDied","Data":"be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f"} Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.274303 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.326295 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-util\") pod \"489dadce-846b-4338-b396-1c80e256b2f8\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.326359 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-bundle\") pod \"489dadce-846b-4338-b396-1c80e256b2f8\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.326444 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdknz\" (UniqueName: \"kubernetes.io/projected/489dadce-846b-4338-b396-1c80e256b2f8-kube-api-access-hdknz\") pod \"489dadce-846b-4338-b396-1c80e256b2f8\" (UID: \"489dadce-846b-4338-b396-1c80e256b2f8\") " Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.327525 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-bundle" (OuterVolumeSpecName: "bundle") pod "489dadce-846b-4338-b396-1c80e256b2f8" (UID: "489dadce-846b-4338-b396-1c80e256b2f8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.334534 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/489dadce-846b-4338-b396-1c80e256b2f8-kube-api-access-hdknz" (OuterVolumeSpecName: "kube-api-access-hdknz") pod "489dadce-846b-4338-b396-1c80e256b2f8" (UID: "489dadce-846b-4338-b396-1c80e256b2f8"). InnerVolumeSpecName "kube-api-access-hdknz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.342032 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-util" (OuterVolumeSpecName: "util") pod "489dadce-846b-4338-b396-1c80e256b2f8" (UID: "489dadce-846b-4338-b396-1c80e256b2f8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.427846 4876 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-util\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.427891 4876 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489dadce-846b-4338-b396-1c80e256b2f8-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.427902 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdknz\" (UniqueName: \"kubernetes.io/projected/489dadce-846b-4338-b396-1c80e256b2f8-kube-api-access-hdknz\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.965832 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfrvt" event={"ID":"29ad6925-f568-4472-a5b9-a2e2e1a135b8","Type":"ContainerStarted","Data":"67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96"} Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.969225 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" event={"ID":"489dadce-846b-4338-b396-1c80e256b2f8","Type":"ContainerDied","Data":"724f7a632e0d857d77b337e4c1305948096297e9939537c25c5ff30165f49965"} Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.969321 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="724f7a632e0d857d77b337e4c1305948096297e9939537c25c5ff30165f49965" Mar 13 07:53:16 crc kubenswrapper[4876]: I0313 07:53:16.969426 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c" Mar 13 07:53:17 crc kubenswrapper[4876]: I0313 07:53:17.008448 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pfrvt" podStartSLOduration=2.564431756 podStartE2EDuration="5.008411298s" podCreationTimestamp="2026-03-13 07:53:12 +0000 UTC" firstStartedPulling="2026-03-13 07:53:13.922037214 +0000 UTC m=+853.592816186" lastFinishedPulling="2026-03-13 07:53:16.366016726 +0000 UTC m=+856.036795728" observedRunningTime="2026-03-13 07:53:17.000041963 +0000 UTC m=+856.670821015" watchObservedRunningTime="2026-03-13 07:53:17.008411298 +0000 UTC m=+856.679190320" Mar 13 07:53:23 crc kubenswrapper[4876]: I0313 07:53:23.326855 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:23 crc kubenswrapper[4876]: I0313 07:53:23.327722 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:24 crc kubenswrapper[4876]: I0313 07:53:24.373877 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pfrvt" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="registry-server" probeResult="failure" output=< Mar 13 07:53:24 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 07:53:24 crc kubenswrapper[4876]: > Mar 13 07:53:24 crc kubenswrapper[4876]: I0313 07:53:24.607303 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:53:24 crc kubenswrapper[4876]: I0313 07:53:24.607396 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:53:24 crc kubenswrapper[4876]: I0313 07:53:24.607507 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:53:24 crc kubenswrapper[4876]: I0313 07:53:24.608654 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ebde89ffdd3ea68ea7599ed04541e15278a60938a622659662a7eee8835605e"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 07:53:24 crc kubenswrapper[4876]: I0313 07:53:24.608724 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://8ebde89ffdd3ea68ea7599ed04541e15278a60938a622659662a7eee8835605e" gracePeriod=600 Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.023823 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="8ebde89ffdd3ea68ea7599ed04541e15278a60938a622659662a7eee8835605e" exitCode=0 Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.023898 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"8ebde89ffdd3ea68ea7599ed04541e15278a60938a622659662a7eee8835605e"} Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.024286 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"82566a3994e302f89937ff7ed8edf05f5c82529ba0688b6946e78bb6b8475402"} Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.024319 4876 scope.go:117] "RemoveContainer" containerID="ba651ad626e0def50adb9d0b266fb9ee781c393b76906dacbf80c73b1a02e496" Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.990630 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58"] Mar 13 07:53:25 crc kubenswrapper[4876]: E0313 07:53:25.993181 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489dadce-846b-4338-b396-1c80e256b2f8" containerName="pull" Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.993293 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="489dadce-846b-4338-b396-1c80e256b2f8" containerName="pull" Mar 13 07:53:25 crc kubenswrapper[4876]: E0313 07:53:25.993386 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489dadce-846b-4338-b396-1c80e256b2f8" containerName="util" Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.993451 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="489dadce-846b-4338-b396-1c80e256b2f8" containerName="util" Mar 13 07:53:25 crc kubenswrapper[4876]: E0313 07:53:25.993528 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489dadce-846b-4338-b396-1c80e256b2f8" containerName="extract" Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.993591 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="489dadce-846b-4338-b396-1c80e256b2f8" containerName="extract" Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.993858 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="489dadce-846b-4338-b396-1c80e256b2f8" containerName="extract" Mar 13 07:53:25 crc kubenswrapper[4876]: I0313 07:53:25.994692 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:25.998601 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:25.999465 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8h7pj" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:25.999602 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:25.999718 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:25.999832 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.088978 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-apiservice-cert\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.089190 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-webhook-cert\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.089316 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgp8s\" (UniqueName: \"kubernetes.io/projected/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-kube-api-access-lgp8s\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.100895 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58"] Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.189997 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgp8s\" (UniqueName: \"kubernetes.io/projected/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-kube-api-access-lgp8s\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.190444 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-apiservice-cert\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.190614 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-webhook-cert\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.201615 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-webhook-cert\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.202405 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-apiservice-cert\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.220414 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgp8s\" (UniqueName: \"kubernetes.io/projected/122ad6c8-564c-4aff-8dc5-6515b0e9b5cb-kube-api-access-lgp8s\") pod \"metallb-operator-controller-manager-5f79f8644-vxb58\" (UID: \"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb\") " pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.313326 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.337561 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn"] Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.338820 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.341175 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.341336 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-6q2h7" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.344467 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.369062 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn"] Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.393874 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26trt\" (UniqueName: \"kubernetes.io/projected/e918dae3-24c9-4e8f-8f89-f4070580eeff-kube-api-access-26trt\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.393925 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e918dae3-24c9-4e8f-8f89-f4070580eeff-webhook-cert\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.393991 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e918dae3-24c9-4e8f-8f89-f4070580eeff-apiservice-cert\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.494850 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e918dae3-24c9-4e8f-8f89-f4070580eeff-apiservice-cert\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.494906 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26trt\" (UniqueName: \"kubernetes.io/projected/e918dae3-24c9-4e8f-8f89-f4070580eeff-kube-api-access-26trt\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.494928 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e918dae3-24c9-4e8f-8f89-f4070580eeff-webhook-cert\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.499424 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e918dae3-24c9-4e8f-8f89-f4070580eeff-webhook-cert\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.513919 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e918dae3-24c9-4e8f-8f89-f4070580eeff-apiservice-cert\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.526798 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26trt\" (UniqueName: \"kubernetes.io/projected/e918dae3-24c9-4e8f-8f89-f4070580eeff-kube-api-access-26trt\") pod \"metallb-operator-webhook-server-679dfbf9b7-f5wpn\" (UID: \"e918dae3-24c9-4e8f-8f89-f4070580eeff\") " pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.677935 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58"] Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.684048 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:26 crc kubenswrapper[4876]: I0313 07:53:26.916681 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn"] Mar 13 07:53:26 crc kubenswrapper[4876]: W0313 07:53:26.925384 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode918dae3_24c9_4e8f_8f89_f4070580eeff.slice/crio-2c0dcb90c49adab4988efbef521b4fbdb92bd2919ab9d37f8cd5bd2843b27026 WatchSource:0}: Error finding container 2c0dcb90c49adab4988efbef521b4fbdb92bd2919ab9d37f8cd5bd2843b27026: Status 404 returned error can't find the container with id 2c0dcb90c49adab4988efbef521b4fbdb92bd2919ab9d37f8cd5bd2843b27026 Mar 13 07:53:27 crc kubenswrapper[4876]: I0313 07:53:27.056170 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" event={"ID":"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb","Type":"ContainerStarted","Data":"ba97ebe4907bd39265aa349e11ca8aaf063c0bcf6f10cd9b64a2b8b396baf4b3"} Mar 13 07:53:27 crc kubenswrapper[4876]: I0313 07:53:27.058525 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" event={"ID":"e918dae3-24c9-4e8f-8f89-f4070580eeff","Type":"ContainerStarted","Data":"2c0dcb90c49adab4988efbef521b4fbdb92bd2919ab9d37f8cd5bd2843b27026"} Mar 13 07:53:33 crc kubenswrapper[4876]: I0313 07:53:33.390436 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:33 crc kubenswrapper[4876]: I0313 07:53:33.447819 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:33 crc kubenswrapper[4876]: I0313 07:53:33.628767 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pfrvt"] Mar 13 07:53:34 crc kubenswrapper[4876]: I0313 07:53:34.132690 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" event={"ID":"e918dae3-24c9-4e8f-8f89-f4070580eeff","Type":"ContainerStarted","Data":"f86280c7261290ebd801614febae9f96a3b98ed5f792cc5d9870b5c176ee773f"} Mar 13 07:53:34 crc kubenswrapper[4876]: I0313 07:53:34.133591 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:53:34 crc kubenswrapper[4876]: I0313 07:53:34.136719 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" event={"ID":"122ad6c8-564c-4aff-8dc5-6515b0e9b5cb","Type":"ContainerStarted","Data":"9f1d57578cb31b36ca9c71c297e38aa0cbb24e6ace3ec099df12a26814a71eb7"} Mar 13 07:53:34 crc kubenswrapper[4876]: I0313 07:53:34.136755 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:53:34 crc kubenswrapper[4876]: I0313 07:53:34.160258 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" podStartSLOduration=1.523144892 podStartE2EDuration="8.160212963s" podCreationTimestamp="2026-03-13 07:53:26 +0000 UTC" firstStartedPulling="2026-03-13 07:53:26.929021802 +0000 UTC m=+866.599800774" lastFinishedPulling="2026-03-13 07:53:33.566089863 +0000 UTC m=+873.236868845" observedRunningTime="2026-03-13 07:53:34.157150606 +0000 UTC m=+873.827929588" watchObservedRunningTime="2026-03-13 07:53:34.160212963 +0000 UTC m=+873.830991945" Mar 13 07:53:34 crc kubenswrapper[4876]: I0313 07:53:34.189365 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" podStartSLOduration=2.329883363 podStartE2EDuration="9.189340135s" podCreationTimestamp="2026-03-13 07:53:25 +0000 UTC" firstStartedPulling="2026-03-13 07:53:26.69810307 +0000 UTC m=+866.368882052" lastFinishedPulling="2026-03-13 07:53:33.557559842 +0000 UTC m=+873.228338824" observedRunningTime="2026-03-13 07:53:34.18386563 +0000 UTC m=+873.854644612" watchObservedRunningTime="2026-03-13 07:53:34.189340135 +0000 UTC m=+873.860119117" Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.142568 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pfrvt" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="registry-server" containerID="cri-o://67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96" gracePeriod=2 Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.515050 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.641942 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-catalog-content\") pod \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.642080 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-utilities\") pod \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.642115 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbpt2\" (UniqueName: \"kubernetes.io/projected/29ad6925-f568-4472-a5b9-a2e2e1a135b8-kube-api-access-tbpt2\") pod \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\" (UID: \"29ad6925-f568-4472-a5b9-a2e2e1a135b8\") " Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.642921 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-utilities" (OuterVolumeSpecName: "utilities") pod "29ad6925-f568-4472-a5b9-a2e2e1a135b8" (UID: "29ad6925-f568-4472-a5b9-a2e2e1a135b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.660955 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29ad6925-f568-4472-a5b9-a2e2e1a135b8-kube-api-access-tbpt2" (OuterVolumeSpecName: "kube-api-access-tbpt2") pod "29ad6925-f568-4472-a5b9-a2e2e1a135b8" (UID: "29ad6925-f568-4472-a5b9-a2e2e1a135b8"). InnerVolumeSpecName "kube-api-access-tbpt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.743575 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.743626 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbpt2\" (UniqueName: \"kubernetes.io/projected/29ad6925-f568-4472-a5b9-a2e2e1a135b8-kube-api-access-tbpt2\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.782609 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29ad6925-f568-4472-a5b9-a2e2e1a135b8" (UID: "29ad6925-f568-4472-a5b9-a2e2e1a135b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:53:35 crc kubenswrapper[4876]: I0313 07:53:35.844843 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29ad6925-f568-4472-a5b9-a2e2e1a135b8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.151176 4876 generic.go:334] "Generic (PLEG): container finished" podID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerID="67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96" exitCode=0 Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.151232 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfrvt" event={"ID":"29ad6925-f568-4472-a5b9-a2e2e1a135b8","Type":"ContainerDied","Data":"67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96"} Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.151278 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfrvt" event={"ID":"29ad6925-f568-4472-a5b9-a2e2e1a135b8","Type":"ContainerDied","Data":"bf63f34186ad1266b9eafe5fe7d122a9d59fc811f3b2b13799372d641266ff9a"} Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.151296 4876 scope.go:117] "RemoveContainer" containerID="67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.151457 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pfrvt" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.173000 4876 scope.go:117] "RemoveContainer" containerID="be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.198092 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pfrvt"] Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.201936 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pfrvt"] Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.216045 4876 scope.go:117] "RemoveContainer" containerID="932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.237338 4876 scope.go:117] "RemoveContainer" containerID="67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96" Mar 13 07:53:36 crc kubenswrapper[4876]: E0313 07:53:36.237852 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96\": container with ID starting with 67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96 not found: ID does not exist" containerID="67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.237891 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96"} err="failed to get container status \"67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96\": rpc error: code = NotFound desc = could not find container \"67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96\": container with ID starting with 67888a443f2ff822b17e75afe8b9ce996a00a13b15e81feea96388f0645a0a96 not found: ID does not exist" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.237918 4876 scope.go:117] "RemoveContainer" containerID="be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f" Mar 13 07:53:36 crc kubenswrapper[4876]: E0313 07:53:36.238176 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f\": container with ID starting with be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f not found: ID does not exist" containerID="be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.238201 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f"} err="failed to get container status \"be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f\": rpc error: code = NotFound desc = could not find container \"be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f\": container with ID starting with be18eac1da5b4850ab2e52ea703fcd7fe880f0275b0effefc286a3d3a7211f6f not found: ID does not exist" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.238219 4876 scope.go:117] "RemoveContainer" containerID="932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686" Mar 13 07:53:36 crc kubenswrapper[4876]: E0313 07:53:36.238938 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686\": container with ID starting with 932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686 not found: ID does not exist" containerID="932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686" Mar 13 07:53:36 crc kubenswrapper[4876]: I0313 07:53:36.238964 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686"} err="failed to get container status \"932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686\": rpc error: code = NotFound desc = could not find container \"932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686\": container with ID starting with 932057cc58807b3c6cd3f8e148b90e5f0941d66426f7461a234cff8f47980686 not found: ID does not exist" Mar 13 07:53:37 crc kubenswrapper[4876]: I0313 07:53:37.043527 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" path="/var/lib/kubelet/pods/29ad6925-f568-4472-a5b9-a2e2e1a135b8/volumes" Mar 13 07:53:46 crc kubenswrapper[4876]: I0313 07:53:46.689930 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-679dfbf9b7-f5wpn" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.139621 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556474-24kvs"] Mar 13 07:54:00 crc kubenswrapper[4876]: E0313 07:54:00.140499 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="extract-utilities" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.140513 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="extract-utilities" Mar 13 07:54:00 crc kubenswrapper[4876]: E0313 07:54:00.140530 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="extract-content" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.140537 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="extract-content" Mar 13 07:54:00 crc kubenswrapper[4876]: E0313 07:54:00.140552 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="registry-server" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.140559 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="registry-server" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.140660 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ad6925-f568-4472-a5b9-a2e2e1a135b8" containerName="registry-server" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.141104 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556474-24kvs" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.143952 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.144003 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.144986 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.158250 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556474-24kvs"] Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.291309 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqvw9\" (UniqueName: \"kubernetes.io/projected/5e2a6f4f-87ef-4109-a53d-c48a82c094aa-kube-api-access-mqvw9\") pod \"auto-csr-approver-29556474-24kvs\" (UID: \"5e2a6f4f-87ef-4109-a53d-c48a82c094aa\") " pod="openshift-infra/auto-csr-approver-29556474-24kvs" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.393060 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqvw9\" (UniqueName: \"kubernetes.io/projected/5e2a6f4f-87ef-4109-a53d-c48a82c094aa-kube-api-access-mqvw9\") pod \"auto-csr-approver-29556474-24kvs\" (UID: \"5e2a6f4f-87ef-4109-a53d-c48a82c094aa\") " pod="openshift-infra/auto-csr-approver-29556474-24kvs" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.414598 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqvw9\" (UniqueName: \"kubernetes.io/projected/5e2a6f4f-87ef-4109-a53d-c48a82c094aa-kube-api-access-mqvw9\") pod \"auto-csr-approver-29556474-24kvs\" (UID: \"5e2a6f4f-87ef-4109-a53d-c48a82c094aa\") " pod="openshift-infra/auto-csr-approver-29556474-24kvs" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.507761 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556474-24kvs" Mar 13 07:54:00 crc kubenswrapper[4876]: I0313 07:54:00.739558 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556474-24kvs"] Mar 13 07:54:01 crc kubenswrapper[4876]: I0313 07:54:01.343330 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556474-24kvs" event={"ID":"5e2a6f4f-87ef-4109-a53d-c48a82c094aa","Type":"ContainerStarted","Data":"6e4aa061fcb7af635e5d4e7ae2556ae6f989cc02e854d4a4c559b2c9099e1099"} Mar 13 07:54:02 crc kubenswrapper[4876]: I0313 07:54:02.355283 4876 generic.go:334] "Generic (PLEG): container finished" podID="5e2a6f4f-87ef-4109-a53d-c48a82c094aa" containerID="4b928f88852dc04e2f1b3cad2b15db0da8b97c2bc616434bde6ec1839f321c5b" exitCode=0 Mar 13 07:54:02 crc kubenswrapper[4876]: I0313 07:54:02.355361 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556474-24kvs" event={"ID":"5e2a6f4f-87ef-4109-a53d-c48a82c094aa","Type":"ContainerDied","Data":"4b928f88852dc04e2f1b3cad2b15db0da8b97c2bc616434bde6ec1839f321c5b"} Mar 13 07:54:03 crc kubenswrapper[4876]: I0313 07:54:03.654979 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556474-24kvs" Mar 13 07:54:03 crc kubenswrapper[4876]: I0313 07:54:03.743923 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqvw9\" (UniqueName: \"kubernetes.io/projected/5e2a6f4f-87ef-4109-a53d-c48a82c094aa-kube-api-access-mqvw9\") pod \"5e2a6f4f-87ef-4109-a53d-c48a82c094aa\" (UID: \"5e2a6f4f-87ef-4109-a53d-c48a82c094aa\") " Mar 13 07:54:03 crc kubenswrapper[4876]: I0313 07:54:03.752313 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2a6f4f-87ef-4109-a53d-c48a82c094aa-kube-api-access-mqvw9" (OuterVolumeSpecName: "kube-api-access-mqvw9") pod "5e2a6f4f-87ef-4109-a53d-c48a82c094aa" (UID: "5e2a6f4f-87ef-4109-a53d-c48a82c094aa"). InnerVolumeSpecName "kube-api-access-mqvw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:54:03 crc kubenswrapper[4876]: I0313 07:54:03.845793 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqvw9\" (UniqueName: \"kubernetes.io/projected/5e2a6f4f-87ef-4109-a53d-c48a82c094aa-kube-api-access-mqvw9\") on node \"crc\" DevicePath \"\"" Mar 13 07:54:04 crc kubenswrapper[4876]: I0313 07:54:04.369745 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556474-24kvs" event={"ID":"5e2a6f4f-87ef-4109-a53d-c48a82c094aa","Type":"ContainerDied","Data":"6e4aa061fcb7af635e5d4e7ae2556ae6f989cc02e854d4a4c559b2c9099e1099"} Mar 13 07:54:04 crc kubenswrapper[4876]: I0313 07:54:04.369798 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e4aa061fcb7af635e5d4e7ae2556ae6f989cc02e854d4a4c559b2c9099e1099" Mar 13 07:54:04 crc kubenswrapper[4876]: I0313 07:54:04.369888 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556474-24kvs" Mar 13 07:54:04 crc kubenswrapper[4876]: I0313 07:54:04.716790 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556468-njzx8"] Mar 13 07:54:04 crc kubenswrapper[4876]: I0313 07:54:04.722199 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556468-njzx8"] Mar 13 07:54:05 crc kubenswrapper[4876]: I0313 07:54:05.047208 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="837c43f4-e275-4668-b5c2-34f7eead6093" path="/var/lib/kubelet/pods/837c43f4-e275-4668-b5c2-34f7eead6093/volumes" Mar 13 07:54:06 crc kubenswrapper[4876]: I0313 07:54:06.316540 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5f79f8644-vxb58" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.212262 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hmszn"] Mar 13 07:54:07 crc kubenswrapper[4876]: E0313 07:54:07.212621 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2a6f4f-87ef-4109-a53d-c48a82c094aa" containerName="oc" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.212636 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2a6f4f-87ef-4109-a53d-c48a82c094aa" containerName="oc" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.212764 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2a6f4f-87ef-4109-a53d-c48a82c094aa" containerName="oc" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.214908 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.217279 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-w7zbl" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.217628 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.224638 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq"] Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.225812 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.227828 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.233311 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.239747 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq"] Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.303700 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-reloader\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.303804 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0979a634-546c-441c-8438-51ab6723d891-metrics-certs\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.303840 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btw5d\" (UniqueName: \"kubernetes.io/projected/0979a634-546c-441c-8438-51ab6723d891-kube-api-access-btw5d\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.304000 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-metrics\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.304101 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-frr-sockets\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.304197 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxfrw\" (UniqueName: \"kubernetes.io/projected/91b6f5ca-3921-446a-a57a-de264ca63118-kube-api-access-zxfrw\") pod \"frr-k8s-webhook-server-bcc4b6f68-6vvzq\" (UID: \"91b6f5ca-3921-446a-a57a-de264ca63118\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.304412 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0979a634-546c-441c-8438-51ab6723d891-frr-startup\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.304486 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91b6f5ca-3921-446a-a57a-de264ca63118-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-6vvzq\" (UID: \"91b6f5ca-3921-446a-a57a-de264ca63118\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.304513 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-frr-conf\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.321589 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-nb74f"] Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.322603 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.327012 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.327210 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-xtfhw" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.329206 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.342716 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.362051 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-bmq26"] Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.385972 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.389731 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.390321 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-bmq26"] Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406395 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-metrics\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406457 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-frr-sockets\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406491 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxfrw\" (UniqueName: \"kubernetes.io/projected/91b6f5ca-3921-446a-a57a-de264ca63118-kube-api-access-zxfrw\") pod \"frr-k8s-webhook-server-bcc4b6f68-6vvzq\" (UID: \"91b6f5ca-3921-446a-a57a-de264ca63118\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406522 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-metrics-certs\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406552 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xrcn\" (UniqueName: \"kubernetes.io/projected/65146c95-889a-4030-b7d3-3a1810b158de-kube-api-access-6xrcn\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406581 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/65146c95-889a-4030-b7d3-3a1810b158de-metallb-excludel2\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406602 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0979a634-546c-441c-8438-51ab6723d891-frr-startup\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406619 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91b6f5ca-3921-446a-a57a-de264ca63118-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-6vvzq\" (UID: \"91b6f5ca-3921-446a-a57a-de264ca63118\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406640 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-frr-conf\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406763 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-reloader\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406851 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0979a634-546c-441c-8438-51ab6723d891-metrics-certs\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406923 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.406947 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btw5d\" (UniqueName: \"kubernetes.io/projected/0979a634-546c-441c-8438-51ab6723d891-kube-api-access-btw5d\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.407181 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-frr-conf\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: E0313 07:54:07.407602 4876 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Mar 13 07:54:07 crc kubenswrapper[4876]: E0313 07:54:07.407702 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0979a634-546c-441c-8438-51ab6723d891-metrics-certs podName:0979a634-546c-441c-8438-51ab6723d891 nodeName:}" failed. No retries permitted until 2026-03-13 07:54:07.907672842 +0000 UTC m=+907.578452024 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0979a634-546c-441c-8438-51ab6723d891-metrics-certs") pod "frr-k8s-hmszn" (UID: "0979a634-546c-441c-8438-51ab6723d891") : secret "frr-k8s-certs-secret" not found Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.407779 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-metrics\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.407946 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-frr-sockets\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.408191 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0979a634-546c-441c-8438-51ab6723d891-reloader\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.414949 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/91b6f5ca-3921-446a-a57a-de264ca63118-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-6vvzq\" (UID: \"91b6f5ca-3921-446a-a57a-de264ca63118\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.417484 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0979a634-546c-441c-8438-51ab6723d891-frr-startup\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.432273 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btw5d\" (UniqueName: \"kubernetes.io/projected/0979a634-546c-441c-8438-51ab6723d891-kube-api-access-btw5d\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.438383 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxfrw\" (UniqueName: \"kubernetes.io/projected/91b6f5ca-3921-446a-a57a-de264ca63118-kube-api-access-zxfrw\") pod \"frr-k8s-webhook-server-bcc4b6f68-6vvzq\" (UID: \"91b6f5ca-3921-446a-a57a-de264ca63118\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.508488 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmqd6\" (UniqueName: \"kubernetes.io/projected/97a22a43-697e-4c6a-b800-489008955cbe-kube-api-access-tmqd6\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.508579 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.508616 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97a22a43-697e-4c6a-b800-489008955cbe-metrics-certs\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.508642 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-metrics-certs\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.508667 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xrcn\" (UniqueName: \"kubernetes.io/projected/65146c95-889a-4030-b7d3-3a1810b158de-kube-api-access-6xrcn\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.508689 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97a22a43-697e-4c6a-b800-489008955cbe-cert\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.508713 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/65146c95-889a-4030-b7d3-3a1810b158de-metallb-excludel2\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: E0313 07:54:07.509350 4876 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 07:54:07 crc kubenswrapper[4876]: E0313 07:54:07.509923 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist podName:65146c95-889a-4030-b7d3-3a1810b158de nodeName:}" failed. No retries permitted until 2026-03-13 07:54:08.009898549 +0000 UTC m=+907.680677521 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist") pod "speaker-nb74f" (UID: "65146c95-889a-4030-b7d3-3a1810b158de") : secret "metallb-memberlist" not found Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.510565 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/65146c95-889a-4030-b7d3-3a1810b158de-metallb-excludel2\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.513679 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-metrics-certs\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.528354 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xrcn\" (UniqueName: \"kubernetes.io/projected/65146c95-889a-4030-b7d3-3a1810b158de-kube-api-access-6xrcn\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.544955 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.610941 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97a22a43-697e-4c6a-b800-489008955cbe-metrics-certs\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.611035 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97a22a43-697e-4c6a-b800-489008955cbe-cert\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.611101 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmqd6\" (UniqueName: \"kubernetes.io/projected/97a22a43-697e-4c6a-b800-489008955cbe-kube-api-access-tmqd6\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.615305 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.622335 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97a22a43-697e-4c6a-b800-489008955cbe-metrics-certs\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.626073 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97a22a43-697e-4c6a-b800-489008955cbe-cert\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.631154 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmqd6\" (UniqueName: \"kubernetes.io/projected/97a22a43-697e-4c6a-b800-489008955cbe-kube-api-access-tmqd6\") pod \"controller-7bb4cc7c98-bmq26\" (UID: \"97a22a43-697e-4c6a-b800-489008955cbe\") " pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.713988 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.778183 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq"] Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.913901 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0979a634-546c-441c-8438-51ab6723d891-metrics-certs\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.918810 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0979a634-546c-441c-8438-51ab6723d891-metrics-certs\") pod \"frr-k8s-hmszn\" (UID: \"0979a634-546c-441c-8438-51ab6723d891\") " pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:07 crc kubenswrapper[4876]: I0313 07:54:07.965402 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-bmq26"] Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.015508 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:08 crc kubenswrapper[4876]: E0313 07:54:08.015737 4876 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 07:54:08 crc kubenswrapper[4876]: E0313 07:54:08.015827 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist podName:65146c95-889a-4030-b7d3-3a1810b158de nodeName:}" failed. No retries permitted until 2026-03-13 07:54:09.015809018 +0000 UTC m=+908.686588000 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist") pod "speaker-nb74f" (UID: "65146c95-889a-4030-b7d3-3a1810b158de") : secret "metallb-memberlist" not found Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.132213 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.425155 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-bmq26" event={"ID":"97a22a43-697e-4c6a-b800-489008955cbe","Type":"ContainerStarted","Data":"961e03b373a77d158500bc7852a55289dd20c0e2fe6dd87d1bb600745aa8e364"} Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.425230 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-bmq26" event={"ID":"97a22a43-697e-4c6a-b800-489008955cbe","Type":"ContainerStarted","Data":"0c0c299fb07a38bd319ba956cdd0b276cc7bd3b8e614f109691fe02a1466ff47"} Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.425267 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-bmq26" event={"ID":"97a22a43-697e-4c6a-b800-489008955cbe","Type":"ContainerStarted","Data":"2b0b688514fb81a89558251331fe12ca40a6d6fc4ef218eb15fc3bf285995a38"} Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.425626 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.431295 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" event={"ID":"91b6f5ca-3921-446a-a57a-de264ca63118","Type":"ContainerStarted","Data":"170ddeab7f2ee9cfefb16ed17ce4d747a110119a8374fef86d4ab7554efcec6c"} Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.433540 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerStarted","Data":"c3aa5d66d35f41dbfb61af8321ae671a471f7c2cf024bf38171611aefdb9615d"} Mar 13 07:54:08 crc kubenswrapper[4876]: I0313 07:54:08.472851 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-bmq26" podStartSLOduration=1.472825064 podStartE2EDuration="1.472825064s" podCreationTimestamp="2026-03-13 07:54:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:54:08.469711466 +0000 UTC m=+908.140490458" watchObservedRunningTime="2026-03-13 07:54:08.472825064 +0000 UTC m=+908.143604036" Mar 13 07:54:09 crc kubenswrapper[4876]: I0313 07:54:09.031000 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:09 crc kubenswrapper[4876]: I0313 07:54:09.037298 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/65146c95-889a-4030-b7d3-3a1810b158de-memberlist\") pod \"speaker-nb74f\" (UID: \"65146c95-889a-4030-b7d3-3a1810b158de\") " pod="metallb-system/speaker-nb74f" Mar 13 07:54:09 crc kubenswrapper[4876]: I0313 07:54:09.139433 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nb74f" Mar 13 07:54:09 crc kubenswrapper[4876]: W0313 07:54:09.165603 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65146c95_889a_4030_b7d3_3a1810b158de.slice/crio-6c3b8ca2bbce8100b0a3c578a7a7e0811e76fbe921bb01f506caba7348fc8977 WatchSource:0}: Error finding container 6c3b8ca2bbce8100b0a3c578a7a7e0811e76fbe921bb01f506caba7348fc8977: Status 404 returned error can't find the container with id 6c3b8ca2bbce8100b0a3c578a7a7e0811e76fbe921bb01f506caba7348fc8977 Mar 13 07:54:09 crc kubenswrapper[4876]: I0313 07:54:09.455995 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nb74f" event={"ID":"65146c95-889a-4030-b7d3-3a1810b158de","Type":"ContainerStarted","Data":"73017e96ab4cf0f1272e80c9bbfb49860ad6717452b45e9e3f607f5a2ac4d265"} Mar 13 07:54:09 crc kubenswrapper[4876]: I0313 07:54:09.456047 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nb74f" event={"ID":"65146c95-889a-4030-b7d3-3a1810b158de","Type":"ContainerStarted","Data":"6c3b8ca2bbce8100b0a3c578a7a7e0811e76fbe921bb01f506caba7348fc8977"} Mar 13 07:54:10 crc kubenswrapper[4876]: I0313 07:54:10.472005 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nb74f" event={"ID":"65146c95-889a-4030-b7d3-3a1810b158de","Type":"ContainerStarted","Data":"ccd3020c3eae9c33cb82912e78a4c9df96ddc50a671fae8519330fbf23671a5a"} Mar 13 07:54:10 crc kubenswrapper[4876]: I0313 07:54:10.472675 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-nb74f" Mar 13 07:54:10 crc kubenswrapper[4876]: I0313 07:54:10.492406 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-nb74f" podStartSLOduration=3.492384169 podStartE2EDuration="3.492384169s" podCreationTimestamp="2026-03-13 07:54:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:54:10.488284033 +0000 UTC m=+910.159063015" watchObservedRunningTime="2026-03-13 07:54:10.492384169 +0000 UTC m=+910.163163151" Mar 13 07:54:16 crc kubenswrapper[4876]: I0313 07:54:16.525488 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" event={"ID":"91b6f5ca-3921-446a-a57a-de264ca63118","Type":"ContainerStarted","Data":"d6c25afa56b095e1a9275e0221b8dd8119ba38627db8435d4d0b7ec6e50a2456"} Mar 13 07:54:16 crc kubenswrapper[4876]: I0313 07:54:16.526368 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:16 crc kubenswrapper[4876]: I0313 07:54:16.529308 4876 generic.go:334] "Generic (PLEG): container finished" podID="0979a634-546c-441c-8438-51ab6723d891" containerID="84daf4981473c6cd91a3a357c5eeedee71aabb47843085f668e852d376946aba" exitCode=0 Mar 13 07:54:16 crc kubenswrapper[4876]: I0313 07:54:16.529342 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerDied","Data":"84daf4981473c6cd91a3a357c5eeedee71aabb47843085f668e852d376946aba"} Mar 13 07:54:16 crc kubenswrapper[4876]: I0313 07:54:16.556483 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" podStartSLOduration=1.724534351 podStartE2EDuration="9.556459168s" podCreationTimestamp="2026-03-13 07:54:07 +0000 UTC" firstStartedPulling="2026-03-13 07:54:07.791505532 +0000 UTC m=+907.462284514" lastFinishedPulling="2026-03-13 07:54:15.623430349 +0000 UTC m=+915.294209331" observedRunningTime="2026-03-13 07:54:16.553366701 +0000 UTC m=+916.224145683" watchObservedRunningTime="2026-03-13 07:54:16.556459168 +0000 UTC m=+916.227238150" Mar 13 07:54:17 crc kubenswrapper[4876]: I0313 07:54:17.540701 4876 generic.go:334] "Generic (PLEG): container finished" podID="0979a634-546c-441c-8438-51ab6723d891" containerID="19db13c3c910c12828922ef6b7f71c5215b62f597b271d435b533431a628707d" exitCode=0 Mar 13 07:54:17 crc kubenswrapper[4876]: I0313 07:54:17.540772 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerDied","Data":"19db13c3c910c12828922ef6b7f71c5215b62f597b271d435b533431a628707d"} Mar 13 07:54:18 crc kubenswrapper[4876]: I0313 07:54:18.553485 4876 generic.go:334] "Generic (PLEG): container finished" podID="0979a634-546c-441c-8438-51ab6723d891" containerID="501302ea24a55e113c2871dcdb4be2fa86dcc3f463c56801c4ee7e643e1198f0" exitCode=0 Mar 13 07:54:18 crc kubenswrapper[4876]: I0313 07:54:18.553546 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerDied","Data":"501302ea24a55e113c2871dcdb4be2fa86dcc3f463c56801c4ee7e643e1198f0"} Mar 13 07:54:19 crc kubenswrapper[4876]: I0313 07:54:19.143843 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-nb74f" Mar 13 07:54:19 crc kubenswrapper[4876]: I0313 07:54:19.566911 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerStarted","Data":"47f9d6e8e5be7629d3963c8dbdc223143d0256c0cdeae1e9a523c3c28e2b3be2"} Mar 13 07:54:19 crc kubenswrapper[4876]: I0313 07:54:19.566987 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerStarted","Data":"767c2c1fe472ad310525a378233a1dd0c7991629b51ed80725fed8253277ad42"} Mar 13 07:54:19 crc kubenswrapper[4876]: I0313 07:54:19.567006 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerStarted","Data":"72d0e8b30d03e82bfbc2a0d75b6ee1adc4af013c70ce7435a670e8b34002ab50"} Mar 13 07:54:19 crc kubenswrapper[4876]: I0313 07:54:19.567023 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerStarted","Data":"5741ef41c8573e320ab0ba4b68f30560c1c29cca3806076b8d85e5a1567f3675"} Mar 13 07:54:20 crc kubenswrapper[4876]: I0313 07:54:20.579887 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerStarted","Data":"8c6f5245607a315ef51eeac6109d9f2f94754ef4adc755f5ad41edf6c6ab994f"} Mar 13 07:54:20 crc kubenswrapper[4876]: I0313 07:54:20.580269 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:20 crc kubenswrapper[4876]: I0313 07:54:20.580287 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hmszn" event={"ID":"0979a634-546c-441c-8438-51ab6723d891","Type":"ContainerStarted","Data":"12132d94d13fcc6f6c03471ef27a3b4457ea819d623ee2e05a1564d1d21bae4a"} Mar 13 07:54:20 crc kubenswrapper[4876]: I0313 07:54:20.603870 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hmszn" podStartSLOduration=6.230367804 podStartE2EDuration="13.603848863s" podCreationTimestamp="2026-03-13 07:54:07 +0000 UTC" firstStartedPulling="2026-03-13 07:54:08.268194095 +0000 UTC m=+907.938973077" lastFinishedPulling="2026-03-13 07:54:15.641675154 +0000 UTC m=+915.312454136" observedRunningTime="2026-03-13 07:54:20.600279473 +0000 UTC m=+920.271058475" watchObservedRunningTime="2026-03-13 07:54:20.603848863 +0000 UTC m=+920.274627845" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.275757 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-94s6n"] Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.277084 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-94s6n" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.327083 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-ptwjl" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.327113 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.327199 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.336328 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-94s6n"] Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.368665 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkjrz\" (UniqueName: \"kubernetes.io/projected/1cd0668d-2ec5-4be4-aeb8-afa043cedf31-kube-api-access-qkjrz\") pod \"openstack-operator-index-94s6n\" (UID: \"1cd0668d-2ec5-4be4-aeb8-afa043cedf31\") " pod="openstack-operators/openstack-operator-index-94s6n" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.470809 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkjrz\" (UniqueName: \"kubernetes.io/projected/1cd0668d-2ec5-4be4-aeb8-afa043cedf31-kube-api-access-qkjrz\") pod \"openstack-operator-index-94s6n\" (UID: \"1cd0668d-2ec5-4be4-aeb8-afa043cedf31\") " pod="openstack-operators/openstack-operator-index-94s6n" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.499768 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkjrz\" (UniqueName: \"kubernetes.io/projected/1cd0668d-2ec5-4be4-aeb8-afa043cedf31-kube-api-access-qkjrz\") pod \"openstack-operator-index-94s6n\" (UID: \"1cd0668d-2ec5-4be4-aeb8-afa043cedf31\") " pod="openstack-operators/openstack-operator-index-94s6n" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.642131 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-94s6n" Mar 13 07:54:22 crc kubenswrapper[4876]: I0313 07:54:22.941437 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-94s6n"] Mar 13 07:54:22 crc kubenswrapper[4876]: W0313 07:54:22.948672 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cd0668d_2ec5_4be4_aeb8_afa043cedf31.slice/crio-81d319fa33dba2fa70c86a71e668ace29f6bf4dd48b9b3937a1bb11093ce425d WatchSource:0}: Error finding container 81d319fa33dba2fa70c86a71e668ace29f6bf4dd48b9b3937a1bb11093ce425d: Status 404 returned error can't find the container with id 81d319fa33dba2fa70c86a71e668ace29f6bf4dd48b9b3937a1bb11093ce425d Mar 13 07:54:23 crc kubenswrapper[4876]: I0313 07:54:23.133324 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:23 crc kubenswrapper[4876]: I0313 07:54:23.180637 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:23 crc kubenswrapper[4876]: I0313 07:54:23.610422 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-94s6n" event={"ID":"1cd0668d-2ec5-4be4-aeb8-afa043cedf31","Type":"ContainerStarted","Data":"81d319fa33dba2fa70c86a71e668ace29f6bf4dd48b9b3937a1bb11093ce425d"} Mar 13 07:54:25 crc kubenswrapper[4876]: I0313 07:54:25.667807 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-94s6n"] Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.265958 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tfbgk"] Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.268460 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.273044 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tfbgk"] Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.331187 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nqwl\" (UniqueName: \"kubernetes.io/projected/ec8ee7ae-49f7-473f-b432-9c2395e93115-kube-api-access-7nqwl\") pod \"openstack-operator-index-tfbgk\" (UID: \"ec8ee7ae-49f7-473f-b432-9c2395e93115\") " pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.432390 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nqwl\" (UniqueName: \"kubernetes.io/projected/ec8ee7ae-49f7-473f-b432-9c2395e93115-kube-api-access-7nqwl\") pod \"openstack-operator-index-tfbgk\" (UID: \"ec8ee7ae-49f7-473f-b432-9c2395e93115\") " pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.455813 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nqwl\" (UniqueName: \"kubernetes.io/projected/ec8ee7ae-49f7-473f-b432-9c2395e93115-kube-api-access-7nqwl\") pod \"openstack-operator-index-tfbgk\" (UID: \"ec8ee7ae-49f7-473f-b432-9c2395e93115\") " pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.594927 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.637500 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-94s6n" event={"ID":"1cd0668d-2ec5-4be4-aeb8-afa043cedf31","Type":"ContainerStarted","Data":"db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749"} Mar 13 07:54:26 crc kubenswrapper[4876]: I0313 07:54:26.676496 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-94s6n" podStartSLOduration=2.151305041 podStartE2EDuration="4.676458874s" podCreationTimestamp="2026-03-13 07:54:22 +0000 UTC" firstStartedPulling="2026-03-13 07:54:22.951148566 +0000 UTC m=+922.621927548" lastFinishedPulling="2026-03-13 07:54:25.476302399 +0000 UTC m=+925.147081381" observedRunningTime="2026-03-13 07:54:26.658101805 +0000 UTC m=+926.328880787" watchObservedRunningTime="2026-03-13 07:54:26.676458874 +0000 UTC m=+926.347237896" Mar 13 07:54:27 crc kubenswrapper[4876]: I0313 07:54:27.067386 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tfbgk"] Mar 13 07:54:27 crc kubenswrapper[4876]: I0313 07:54:27.552892 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-6vvzq" Mar 13 07:54:27 crc kubenswrapper[4876]: I0313 07:54:27.648402 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tfbgk" event={"ID":"ec8ee7ae-49f7-473f-b432-9c2395e93115","Type":"ContainerStarted","Data":"cf1622b60ab6d9d4dec9fbcdca9ef2502aa23ad7bb1d81135d71db459adcfd12"} Mar 13 07:54:27 crc kubenswrapper[4876]: I0313 07:54:27.648473 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tfbgk" event={"ID":"ec8ee7ae-49f7-473f-b432-9c2395e93115","Type":"ContainerStarted","Data":"af516ebf8acec1f2d4f58cc8243c75c64aa24d7d139ae5573564c2bd511f7d64"} Mar 13 07:54:27 crc kubenswrapper[4876]: I0313 07:54:27.648570 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-94s6n" podUID="1cd0668d-2ec5-4be4-aeb8-afa043cedf31" containerName="registry-server" containerID="cri-o://db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749" gracePeriod=2 Mar 13 07:54:27 crc kubenswrapper[4876]: I0313 07:54:27.671892 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tfbgk" podStartSLOduration=1.619087685 podStartE2EDuration="1.671861405s" podCreationTimestamp="2026-03-13 07:54:26 +0000 UTC" firstStartedPulling="2026-03-13 07:54:27.076156412 +0000 UTC m=+926.746935404" lastFinishedPulling="2026-03-13 07:54:27.128930142 +0000 UTC m=+926.799709124" observedRunningTime="2026-03-13 07:54:27.67062117 +0000 UTC m=+927.341400152" watchObservedRunningTime="2026-03-13 07:54:27.671861405 +0000 UTC m=+927.342640397" Mar 13 07:54:27 crc kubenswrapper[4876]: I0313 07:54:27.720433 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-bmq26" Mar 13 07:54:27 crc kubenswrapper[4876]: I0313 07:54:27.996341 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-94s6n" Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.061205 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkjrz\" (UniqueName: \"kubernetes.io/projected/1cd0668d-2ec5-4be4-aeb8-afa043cedf31-kube-api-access-qkjrz\") pod \"1cd0668d-2ec5-4be4-aeb8-afa043cedf31\" (UID: \"1cd0668d-2ec5-4be4-aeb8-afa043cedf31\") " Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.068923 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cd0668d-2ec5-4be4-aeb8-afa043cedf31-kube-api-access-qkjrz" (OuterVolumeSpecName: "kube-api-access-qkjrz") pod "1cd0668d-2ec5-4be4-aeb8-afa043cedf31" (UID: "1cd0668d-2ec5-4be4-aeb8-afa043cedf31"). InnerVolumeSpecName "kube-api-access-qkjrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.137003 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hmszn" Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.168168 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkjrz\" (UniqueName: \"kubernetes.io/projected/1cd0668d-2ec5-4be4-aeb8-afa043cedf31-kube-api-access-qkjrz\") on node \"crc\" DevicePath \"\"" Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.659475 4876 generic.go:334] "Generic (PLEG): container finished" podID="1cd0668d-2ec5-4be4-aeb8-afa043cedf31" containerID="db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749" exitCode=0 Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.659565 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-94s6n" Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.659626 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-94s6n" event={"ID":"1cd0668d-2ec5-4be4-aeb8-afa043cedf31","Type":"ContainerDied","Data":"db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749"} Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.659674 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-94s6n" event={"ID":"1cd0668d-2ec5-4be4-aeb8-afa043cedf31","Type":"ContainerDied","Data":"81d319fa33dba2fa70c86a71e668ace29f6bf4dd48b9b3937a1bb11093ce425d"} Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.659697 4876 scope.go:117] "RemoveContainer" containerID="db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749" Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.686885 4876 scope.go:117] "RemoveContainer" containerID="db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749" Mar 13 07:54:28 crc kubenswrapper[4876]: E0313 07:54:28.687949 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749\": container with ID starting with db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749 not found: ID does not exist" containerID="db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749" Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.688022 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749"} err="failed to get container status \"db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749\": rpc error: code = NotFound desc = could not find container \"db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749\": container with ID starting with db8e729c1e24cc572438399731289f1427c05f3c7b275227724751b177300749 not found: ID does not exist" Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.706942 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-94s6n"] Mar 13 07:54:28 crc kubenswrapper[4876]: I0313 07:54:28.714736 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-94s6n"] Mar 13 07:54:29 crc kubenswrapper[4876]: I0313 07:54:29.051205 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cd0668d-2ec5-4be4-aeb8-afa043cedf31" path="/var/lib/kubelet/pods/1cd0668d-2ec5-4be4-aeb8-afa043cedf31/volumes" Mar 13 07:54:36 crc kubenswrapper[4876]: I0313 07:54:36.596392 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:36 crc kubenswrapper[4876]: I0313 07:54:36.596936 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:36 crc kubenswrapper[4876]: I0313 07:54:36.630945 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:36 crc kubenswrapper[4876]: I0313 07:54:36.761522 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-tfbgk" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.715753 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x"] Mar 13 07:54:39 crc kubenswrapper[4876]: E0313 07:54:39.716687 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd0668d-2ec5-4be4-aeb8-afa043cedf31" containerName="registry-server" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.716708 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd0668d-2ec5-4be4-aeb8-afa043cedf31" containerName="registry-server" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.716903 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd0668d-2ec5-4be4-aeb8-afa043cedf31" containerName="registry-server" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.718511 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.722103 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-zxljl" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.733032 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x"] Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.861474 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-util\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.861609 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-bundle\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.861670 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvlnx\" (UniqueName: \"kubernetes.io/projected/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-kube-api-access-lvlnx\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.964622 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-util\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.964828 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-bundle\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.964912 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvlnx\" (UniqueName: \"kubernetes.io/projected/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-kube-api-access-lvlnx\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.965468 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-util\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.965668 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-bundle\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:39 crc kubenswrapper[4876]: I0313 07:54:39.998437 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvlnx\" (UniqueName: \"kubernetes.io/projected/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-kube-api-access-lvlnx\") pod \"c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:40 crc kubenswrapper[4876]: I0313 07:54:40.055402 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:40 crc kubenswrapper[4876]: I0313 07:54:40.360396 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x"] Mar 13 07:54:40 crc kubenswrapper[4876]: I0313 07:54:40.769955 4876 generic.go:334] "Generic (PLEG): container finished" podID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerID="32182d8970bc27ca2351ddc7af9508181deb90fb8b6b186e7ac4c6d22a3bc957" exitCode=0 Mar 13 07:54:40 crc kubenswrapper[4876]: I0313 07:54:40.771340 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" event={"ID":"28ee98e2-f56e-4050-ae8a-6c4fe84eda60","Type":"ContainerDied","Data":"32182d8970bc27ca2351ddc7af9508181deb90fb8b6b186e7ac4c6d22a3bc957"} Mar 13 07:54:40 crc kubenswrapper[4876]: I0313 07:54:40.771394 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" event={"ID":"28ee98e2-f56e-4050-ae8a-6c4fe84eda60","Type":"ContainerStarted","Data":"525e23e0032d170b6b5f62b41fcc9b144220b9a73323259e50233ac029e454fa"} Mar 13 07:54:41 crc kubenswrapper[4876]: I0313 07:54:41.778602 4876 generic.go:334] "Generic (PLEG): container finished" podID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerID="6172a1a1fd5e991d0f66b3cad7da521d1eb9df39924f479122d40695f9d18f4f" exitCode=0 Mar 13 07:54:41 crc kubenswrapper[4876]: I0313 07:54:41.779668 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" event={"ID":"28ee98e2-f56e-4050-ae8a-6c4fe84eda60","Type":"ContainerDied","Data":"6172a1a1fd5e991d0f66b3cad7da521d1eb9df39924f479122d40695f9d18f4f"} Mar 13 07:54:42 crc kubenswrapper[4876]: I0313 07:54:42.790999 4876 generic.go:334] "Generic (PLEG): container finished" podID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerID="71e9e7934815d7f5f0fe16a2043f2806add72919d5a172706d5a0d4dd2f3276f" exitCode=0 Mar 13 07:54:42 crc kubenswrapper[4876]: I0313 07:54:42.791143 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" event={"ID":"28ee98e2-f56e-4050-ae8a-6c4fe84eda60","Type":"ContainerDied","Data":"71e9e7934815d7f5f0fe16a2043f2806add72919d5a172706d5a0d4dd2f3276f"} Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.193873 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.357846 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-bundle\") pod \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.357990 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvlnx\" (UniqueName: \"kubernetes.io/projected/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-kube-api-access-lvlnx\") pod \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.358103 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-util\") pod \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\" (UID: \"28ee98e2-f56e-4050-ae8a-6c4fe84eda60\") " Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.358603 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-bundle" (OuterVolumeSpecName: "bundle") pod "28ee98e2-f56e-4050-ae8a-6c4fe84eda60" (UID: "28ee98e2-f56e-4050-ae8a-6c4fe84eda60"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.364764 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-kube-api-access-lvlnx" (OuterVolumeSpecName: "kube-api-access-lvlnx") pod "28ee98e2-f56e-4050-ae8a-6c4fe84eda60" (UID: "28ee98e2-f56e-4050-ae8a-6c4fe84eda60"). InnerVolumeSpecName "kube-api-access-lvlnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.373805 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-util" (OuterVolumeSpecName: "util") pod "28ee98e2-f56e-4050-ae8a-6c4fe84eda60" (UID: "28ee98e2-f56e-4050-ae8a-6c4fe84eda60"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.459762 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvlnx\" (UniqueName: \"kubernetes.io/projected/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-kube-api-access-lvlnx\") on node \"crc\" DevicePath \"\"" Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.459795 4876 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-util\") on node \"crc\" DevicePath \"\"" Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.459820 4876 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28ee98e2-f56e-4050-ae8a-6c4fe84eda60-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.809196 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" event={"ID":"28ee98e2-f56e-4050-ae8a-6c4fe84eda60","Type":"ContainerDied","Data":"525e23e0032d170b6b5f62b41fcc9b144220b9a73323259e50233ac029e454fa"} Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.809266 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="525e23e0032d170b6b5f62b41fcc9b144220b9a73323259e50233ac029e454fa" Mar 13 07:54:44 crc kubenswrapper[4876]: I0313 07:54:44.809320 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.650319 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7"] Mar 13 07:54:51 crc kubenswrapper[4876]: E0313 07:54:51.651415 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerName="extract" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.651430 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerName="extract" Mar 13 07:54:51 crc kubenswrapper[4876]: E0313 07:54:51.651443 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerName="pull" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.651450 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerName="pull" Mar 13 07:54:51 crc kubenswrapper[4876]: E0313 07:54:51.651467 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerName="util" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.651500 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerName="util" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.651616 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="28ee98e2-f56e-4050-ae8a-6c4fe84eda60" containerName="extract" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.652138 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.655634 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-r78vj" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.680510 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7"] Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.775145 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktzs6\" (UniqueName: \"kubernetes.io/projected/995e15f2-9906-40df-a498-832366249bf9-kube-api-access-ktzs6\") pod \"openstack-operator-controller-init-578cd8b599-6vlp7\" (UID: \"995e15f2-9906-40df-a498-832366249bf9\") " pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.876448 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktzs6\" (UniqueName: \"kubernetes.io/projected/995e15f2-9906-40df-a498-832366249bf9-kube-api-access-ktzs6\") pod \"openstack-operator-controller-init-578cd8b599-6vlp7\" (UID: \"995e15f2-9906-40df-a498-832366249bf9\") " pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.899068 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktzs6\" (UniqueName: \"kubernetes.io/projected/995e15f2-9906-40df-a498-832366249bf9-kube-api-access-ktzs6\") pod \"openstack-operator-controller-init-578cd8b599-6vlp7\" (UID: \"995e15f2-9906-40df-a498-832366249bf9\") " pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" Mar 13 07:54:51 crc kubenswrapper[4876]: I0313 07:54:51.972312 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" Mar 13 07:54:52 crc kubenswrapper[4876]: I0313 07:54:52.453090 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7"] Mar 13 07:54:52 crc kubenswrapper[4876]: I0313 07:54:52.866835 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" event={"ID":"995e15f2-9906-40df-a498-832366249bf9","Type":"ContainerStarted","Data":"f82bc51a700d03a5af6d63783f3b651b4749b0bd1587dcb1380df52025e00d7f"} Mar 13 07:54:58 crc kubenswrapper[4876]: I0313 07:54:58.920880 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" event={"ID":"995e15f2-9906-40df-a498-832366249bf9","Type":"ContainerStarted","Data":"d0fbd6275d35a013cdaffbea6b6004b696918960c8cf3a6ea5d7fe169b812d7e"} Mar 13 07:54:58 crc kubenswrapper[4876]: I0313 07:54:58.921793 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" Mar 13 07:54:58 crc kubenswrapper[4876]: I0313 07:54:58.972422 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" podStartSLOduration=2.6559068740000003 podStartE2EDuration="7.97239279s" podCreationTimestamp="2026-03-13 07:54:51 +0000 UTC" firstStartedPulling="2026-03-13 07:54:52.468744436 +0000 UTC m=+952.139523418" lastFinishedPulling="2026-03-13 07:54:57.785230352 +0000 UTC m=+957.456009334" observedRunningTime="2026-03-13 07:54:58.966139204 +0000 UTC m=+958.636918186" watchObservedRunningTime="2026-03-13 07:54:58.97239279 +0000 UTC m=+958.643171772" Mar 13 07:55:01 crc kubenswrapper[4876]: I0313 07:55:01.750140 4876 scope.go:117] "RemoveContainer" containerID="1f45548d006a179e52d206bfa29a1c4a272b10d2ec033b3ccd7d01bb5796082a" Mar 13 07:55:11 crc kubenswrapper[4876]: I0313 07:55:11.977984 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-578cd8b599-6vlp7" Mar 13 07:55:24 crc kubenswrapper[4876]: I0313 07:55:24.607793 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:55:24 crc kubenswrapper[4876]: I0313 07:55:24.608558 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.601406 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m6bj8"] Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.603542 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.618187 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6bj8"] Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.789018 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-catalog-content\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.789577 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgfmt\" (UniqueName: \"kubernetes.io/projected/74fa4506-343c-48a2-bb72-612cc2315daa-kube-api-access-tgfmt\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.789654 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-utilities\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.890872 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-utilities\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.891263 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-catalog-content\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.891430 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-utilities\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.891438 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgfmt\" (UniqueName: \"kubernetes.io/projected/74fa4506-343c-48a2-bb72-612cc2315daa-kube-api-access-tgfmt\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.891675 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-catalog-content\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.932415 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgfmt\" (UniqueName: \"kubernetes.io/projected/74fa4506-343c-48a2-bb72-612cc2315daa-kube-api-access-tgfmt\") pod \"redhat-marketplace-m6bj8\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:31 crc kubenswrapper[4876]: I0313 07:55:31.944420 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.283416 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6bj8"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.504946 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.505967 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.510274 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5gnbh" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.521623 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.579436 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.581795 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.587913 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-drvvc" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.597859 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.599720 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.603349 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-25rnf" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.607995 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.640815 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.689739 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.690731 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.694458 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-d59pn" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.708930 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.709760 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xptqz\" (UniqueName: \"kubernetes.io/projected/8d21a99d-11a5-4b5f-a44d-3f9cce84b82a-kube-api-access-xptqz\") pod \"cinder-operator-controller-manager-984cd4dcf-w45l7\" (UID: \"8d21a99d-11a5-4b5f-a44d-3f9cce84b82a\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.709974 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.710437 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hztnv\" (UniqueName: \"kubernetes.io/projected/4cfaf185-ed2d-4c36-9d38-fcf4e74905ec-kube-api-access-hztnv\") pod \"barbican-operator-controller-manager-677bd678f7-6slg2\" (UID: \"4cfaf185-ed2d-4c36-9d38-fcf4e74905ec\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.710504 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9mql\" (UniqueName: \"kubernetes.io/projected/68223df4-94ca-47fe-afbc-93c77725ba2a-kube-api-access-g9mql\") pod \"designate-operator-controller-manager-66d56f6ff4-897gd\" (UID: \"68223df4-94ca-47fe-afbc-93c77725ba2a\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.716009 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-8ts6f" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.742442 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.762528 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.773492 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.774182 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.774736 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.782834 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.783044 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-tt8lc" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.783220 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-p8d4v" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.796082 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.796980 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.802099 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-wxmbl" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.819096 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hztnv\" (UniqueName: \"kubernetes.io/projected/4cfaf185-ed2d-4c36-9d38-fcf4e74905ec-kube-api-access-hztnv\") pod \"barbican-operator-controller-manager-677bd678f7-6slg2\" (UID: \"4cfaf185-ed2d-4c36-9d38-fcf4e74905ec\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.819771 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9mql\" (UniqueName: \"kubernetes.io/projected/68223df4-94ca-47fe-afbc-93c77725ba2a-kube-api-access-g9mql\") pod \"designate-operator-controller-manager-66d56f6ff4-897gd\" (UID: \"68223df4-94ca-47fe-afbc-93c77725ba2a\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.819828 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfz4j\" (UniqueName: \"kubernetes.io/projected/8c04b15f-3ca3-4875-9ea4-45244808af5f-kube-api-access-dfz4j\") pod \"heat-operator-controller-manager-77b6666d85-jcjfd\" (UID: \"8c04b15f-3ca3-4875-9ea4-45244808af5f\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.819859 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96sxz\" (UniqueName: \"kubernetes.io/projected/151ed757-e99e-4e7e-bfaa-7dec0ef8b038-kube-api-access-96sxz\") pod \"glance-operator-controller-manager-5964f64c48-zlt8c\" (UID: \"151ed757-e99e-4e7e-bfaa-7dec0ef8b038\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.819883 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xptqz\" (UniqueName: \"kubernetes.io/projected/8d21a99d-11a5-4b5f-a44d-3f9cce84b82a-kube-api-access-xptqz\") pod \"cinder-operator-controller-manager-984cd4dcf-w45l7\" (UID: \"8d21a99d-11a5-4b5f-a44d-3f9cce84b82a\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.825844 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.828649 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.839860 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.840847 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.853138 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-lrf9n" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.853438 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-5wx5f" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.854879 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.855956 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.864689 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-p69t5" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.865982 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9mql\" (UniqueName: \"kubernetes.io/projected/68223df4-94ca-47fe-afbc-93c77725ba2a-kube-api-access-g9mql\") pod \"designate-operator-controller-manager-66d56f6ff4-897gd\" (UID: \"68223df4-94ca-47fe-afbc-93c77725ba2a\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.869339 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.870215 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.870665 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hztnv\" (UniqueName: \"kubernetes.io/projected/4cfaf185-ed2d-4c36-9d38-fcf4e74905ec-kube-api-access-hztnv\") pod \"barbican-operator-controller-manager-677bd678f7-6slg2\" (UID: \"4cfaf185-ed2d-4c36-9d38-fcf4e74905ec\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.871915 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xptqz\" (UniqueName: \"kubernetes.io/projected/8d21a99d-11a5-4b5f-a44d-3f9cce84b82a-kube-api-access-xptqz\") pod \"cinder-operator-controller-manager-984cd4dcf-w45l7\" (UID: \"8d21a99d-11a5-4b5f-a44d-3f9cce84b82a\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.877749 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-rkxh5" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.896833 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.898059 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.908477 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-c5bf6" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.920013 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921316 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nq9r\" (UniqueName: \"kubernetes.io/projected/98ac6099-e731-43c8-8537-8095f1bfce09-kube-api-access-5nq9r\") pod \"manila-operator-controller-manager-68f45f9d9f-bp9bt\" (UID: \"98ac6099-e731-43c8-8537-8095f1bfce09\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921360 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96sxz\" (UniqueName: \"kubernetes.io/projected/151ed757-e99e-4e7e-bfaa-7dec0ef8b038-kube-api-access-96sxz\") pod \"glance-operator-controller-manager-5964f64c48-zlt8c\" (UID: \"151ed757-e99e-4e7e-bfaa-7dec0ef8b038\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921397 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vvf7\" (UniqueName: \"kubernetes.io/projected/3ff4d1b1-aae8-4900-a317-164ea72a78b2-kube-api-access-8vvf7\") pod \"ironic-operator-controller-manager-6bbb499bbc-n5stb\" (UID: \"3ff4d1b1-aae8-4900-a317-164ea72a78b2\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921428 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921447 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwtgm\" (UniqueName: \"kubernetes.io/projected/33e3356e-590a-4b31-990f-4ecda38efb0a-kube-api-access-rwtgm\") pod \"keystone-operator-controller-manager-684f77d66d-kmhm8\" (UID: \"33e3356e-590a-4b31-990f-4ecda38efb0a\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921466 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzb5q\" (UniqueName: \"kubernetes.io/projected/ff48d75e-19f2-4a48-bcd6-3855b494fdf1-kube-api-access-wzb5q\") pod \"horizon-operator-controller-manager-6d9d6b584d-c6bfj\" (UID: \"ff48d75e-19f2-4a48-bcd6-3855b494fdf1\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921491 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27v49\" (UniqueName: \"kubernetes.io/projected/6aeda87b-2ede-4dcb-9234-bfc8837f4ea8-kube-api-access-27v49\") pod \"mariadb-operator-controller-manager-658d4cdd5-v7kpw\" (UID: \"6aeda87b-2ede-4dcb-9234-bfc8837f4ea8\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921513 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc269\" (UniqueName: \"kubernetes.io/projected/722c5d70-ed42-4b75-8993-3b3643d44d1a-kube-api-access-sc269\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921532 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9k88\" (UniqueName: \"kubernetes.io/projected/ef934388-2da5-47fe-8679-ec520946de1b-kube-api-access-x9k88\") pod \"nova-operator-controller-manager-569cc54c5-pfswh\" (UID: \"ef934388-2da5-47fe-8679-ec520946de1b\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.921556 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb8tt\" (UniqueName: \"kubernetes.io/projected/4284fe2d-d1bb-49f6-b555-0f46419ad11b-kube-api-access-vb8tt\") pod \"neutron-operator-controller-manager-776c5696bf-lnb48\" (UID: \"4284fe2d-d1bb-49f6-b555-0f46419ad11b\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.922041 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfz4j\" (UniqueName: \"kubernetes.io/projected/8c04b15f-3ca3-4875-9ea4-45244808af5f-kube-api-access-dfz4j\") pod \"heat-operator-controller-manager-77b6666d85-jcjfd\" (UID: \"8c04b15f-3ca3-4875-9ea4-45244808af5f\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.931569 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.937623 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-svrsh" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.937844 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.940424 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.951376 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.960458 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw"] Mar 13 07:55:32 crc kubenswrapper[4876]: I0313 07:55:32.991374 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.009338 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.036598 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.038987 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vvf7\" (UniqueName: \"kubernetes.io/projected/3ff4d1b1-aae8-4900-a317-164ea72a78b2-kube-api-access-8vvf7\") pod \"ironic-operator-controller-manager-6bbb499bbc-n5stb\" (UID: \"3ff4d1b1-aae8-4900-a317-164ea72a78b2\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.039056 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.039077 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwtgm\" (UniqueName: \"kubernetes.io/projected/33e3356e-590a-4b31-990f-4ecda38efb0a-kube-api-access-rwtgm\") pod \"keystone-operator-controller-manager-684f77d66d-kmhm8\" (UID: \"33e3356e-590a-4b31-990f-4ecda38efb0a\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.039094 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzb5q\" (UniqueName: \"kubernetes.io/projected/ff48d75e-19f2-4a48-bcd6-3855b494fdf1-kube-api-access-wzb5q\") pod \"horizon-operator-controller-manager-6d9d6b584d-c6bfj\" (UID: \"ff48d75e-19f2-4a48-bcd6-3855b494fdf1\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.039124 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27v49\" (UniqueName: \"kubernetes.io/projected/6aeda87b-2ede-4dcb-9234-bfc8837f4ea8-kube-api-access-27v49\") pod \"mariadb-operator-controller-manager-658d4cdd5-v7kpw\" (UID: \"6aeda87b-2ede-4dcb-9234-bfc8837f4ea8\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.039145 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc269\" (UniqueName: \"kubernetes.io/projected/722c5d70-ed42-4b75-8993-3b3643d44d1a-kube-api-access-sc269\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.039163 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9k88\" (UniqueName: \"kubernetes.io/projected/ef934388-2da5-47fe-8679-ec520946de1b-kube-api-access-x9k88\") pod \"nova-operator-controller-manager-569cc54c5-pfswh\" (UID: \"ef934388-2da5-47fe-8679-ec520946de1b\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.039185 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb8tt\" (UniqueName: \"kubernetes.io/projected/4284fe2d-d1bb-49f6-b555-0f46419ad11b-kube-api-access-vb8tt\") pod \"neutron-operator-controller-manager-776c5696bf-lnb48\" (UID: \"4284fe2d-d1bb-49f6-b555-0f46419ad11b\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.039228 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nq9r\" (UniqueName: \"kubernetes.io/projected/98ac6099-e731-43c8-8537-8095f1bfce09-kube-api-access-5nq9r\") pod \"manila-operator-controller-manager-68f45f9d9f-bp9bt\" (UID: \"98ac6099-e731-43c8-8537-8095f1bfce09\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.039794 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.039846 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert podName:722c5d70-ed42-4b75-8993-3b3643d44d1a nodeName:}" failed. No retries permitted until 2026-03-13 07:55:33.539825905 +0000 UTC m=+993.210604877 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert") pod "infra-operator-controller-manager-5995f4446f-lcwhv" (UID: "722c5d70-ed42-4b75-8993-3b3643d44d1a") : secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.048128 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfz4j\" (UniqueName: \"kubernetes.io/projected/8c04b15f-3ca3-4875-9ea4-45244808af5f-kube-api-access-dfz4j\") pod \"heat-operator-controller-manager-77b6666d85-jcjfd\" (UID: \"8c04b15f-3ca3-4875-9ea4-45244808af5f\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.048705 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96sxz\" (UniqueName: \"kubernetes.io/projected/151ed757-e99e-4e7e-bfaa-7dec0ef8b038-kube-api-access-96sxz\") pod \"glance-operator-controller-manager-5964f64c48-zlt8c\" (UID: \"151ed757-e99e-4e7e-bfaa-7dec0ef8b038\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.075793 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.086061 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.086108 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.089910 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.102309 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.103370 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.107946 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.109142 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.117758 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzb5q\" (UniqueName: \"kubernetes.io/projected/ff48d75e-19f2-4a48-bcd6-3855b494fdf1-kube-api-access-wzb5q\") pod \"horizon-operator-controller-manager-6d9d6b584d-c6bfj\" (UID: \"ff48d75e-19f2-4a48-bcd6-3855b494fdf1\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.118333 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-btq45" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.123870 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vvf7\" (UniqueName: \"kubernetes.io/projected/3ff4d1b1-aae8-4900-a317-164ea72a78b2-kube-api-access-8vvf7\") pod \"ironic-operator-controller-manager-6bbb499bbc-n5stb\" (UID: \"3ff4d1b1-aae8-4900-a317-164ea72a78b2\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.127316 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.128473 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.134680 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-85bf7" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.141057 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkknt\" (UniqueName: \"kubernetes.io/projected/3de3f00b-74de-4328-86e9-9f86c7277deb-kube-api-access-bkknt\") pod \"octavia-operator-controller-manager-5f4f55cb5c-m2xqm\" (UID: \"3de3f00b-74de-4328-86e9-9f86c7277deb\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.149132 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.156641 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.163112 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwtgm\" (UniqueName: \"kubernetes.io/projected/33e3356e-590a-4b31-990f-4ecda38efb0a-kube-api-access-rwtgm\") pod \"keystone-operator-controller-manager-684f77d66d-kmhm8\" (UID: \"33e3356e-590a-4b31-990f-4ecda38efb0a\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.180007 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc269\" (UniqueName: \"kubernetes.io/projected/722c5d70-ed42-4b75-8993-3b3643d44d1a-kube-api-access-sc269\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.180017 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb8tt\" (UniqueName: \"kubernetes.io/projected/4284fe2d-d1bb-49f6-b555-0f46419ad11b-kube-api-access-vb8tt\") pod \"neutron-operator-controller-manager-776c5696bf-lnb48\" (UID: \"4284fe2d-d1bb-49f6-b555-0f46419ad11b\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.180279 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.201612 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nq9r\" (UniqueName: \"kubernetes.io/projected/98ac6099-e731-43c8-8537-8095f1bfce09-kube-api-access-5nq9r\") pod \"manila-operator-controller-manager-68f45f9d9f-bp9bt\" (UID: \"98ac6099-e731-43c8-8537-8095f1bfce09\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.201779 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9k88\" (UniqueName: \"kubernetes.io/projected/ef934388-2da5-47fe-8679-ec520946de1b-kube-api-access-x9k88\") pod \"nova-operator-controller-manager-569cc54c5-pfswh\" (UID: \"ef934388-2da5-47fe-8679-ec520946de1b\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.201883 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27v49\" (UniqueName: \"kubernetes.io/projected/6aeda87b-2ede-4dcb-9234-bfc8837f4ea8-kube-api-access-27v49\") pod \"mariadb-operator-controller-manager-658d4cdd5-v7kpw\" (UID: \"6aeda87b-2ede-4dcb-9234-bfc8837f4ea8\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.210060 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.215421 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.217858 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.224713 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.240885 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-689lm" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.244478 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkknt\" (UniqueName: \"kubernetes.io/projected/3de3f00b-74de-4328-86e9-9f86c7277deb-kube-api-access-bkknt\") pod \"octavia-operator-controller-manager-5f4f55cb5c-m2xqm\" (UID: \"3de3f00b-74de-4328-86e9-9f86c7277deb\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.244543 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.244616 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fvrg\" (UniqueName: \"kubernetes.io/projected/256daf30-6602-4a5d-8a4b-0f96fb5450b2-kube-api-access-2fvrg\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.244741 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbltc\" (UniqueName: \"kubernetes.io/projected/dacb8b62-d4f3-43af-88b2-26fbc725f50f-kube-api-access-tbltc\") pod \"ovn-operator-controller-manager-bbc5b68f9-d5d5r\" (UID: \"dacb8b62-d4f3-43af-88b2-26fbc725f50f\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.250568 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.259041 4876 generic.go:334] "Generic (PLEG): container finished" podID="74fa4506-343c-48a2-bb72-612cc2315daa" containerID="6e548ddbf59d8c444facd3d827172806bf5e476b8006494b502cf04db3c1b714" exitCode=0 Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.260890 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.260959 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6bj8" event={"ID":"74fa4506-343c-48a2-bb72-612cc2315daa","Type":"ContainerDied","Data":"6e548ddbf59d8c444facd3d827172806bf5e476b8006494b502cf04db3c1b714"} Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.261027 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6bj8" event={"ID":"74fa4506-343c-48a2-bb72-612cc2315daa","Type":"ContainerStarted","Data":"b2af2e3544a602c1d17a4e101613d26e8ba41b12441ff2d32d2f34bf72e6f574"} Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.274413 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.328814 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.337025 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.347569 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.347624 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fvrg\" (UniqueName: \"kubernetes.io/projected/256daf30-6602-4a5d-8a4b-0f96fb5450b2-kube-api-access-2fvrg\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.347693 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn6zg\" (UniqueName: \"kubernetes.io/projected/a9a4fe04-37f1-47ad-ab32-e983f13df515-kube-api-access-hn6zg\") pod \"placement-operator-controller-manager-574d45c66c-m7zdz\" (UID: \"a9a4fe04-37f1-47ad-ab32-e983f13df515\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.347710 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbltc\" (UniqueName: \"kubernetes.io/projected/dacb8b62-d4f3-43af-88b2-26fbc725f50f-kube-api-access-tbltc\") pod \"ovn-operator-controller-manager-bbc5b68f9-d5d5r\" (UID: \"dacb8b62-d4f3-43af-88b2-26fbc725f50f\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.348155 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.348211 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert podName:256daf30-6602-4a5d-8a4b-0f96fb5450b2 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:33.848193359 +0000 UTC m=+993.518972341 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" (UID: "256daf30-6602-4a5d-8a4b-0f96fb5450b2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.363163 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkknt\" (UniqueName: \"kubernetes.io/projected/3de3f00b-74de-4328-86e9-9f86c7277deb-kube-api-access-bkknt\") pod \"octavia-operator-controller-manager-5f4f55cb5c-m2xqm\" (UID: \"3de3f00b-74de-4328-86e9-9f86c7277deb\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.414373 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.424848 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.432843 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbltc\" (UniqueName: \"kubernetes.io/projected/dacb8b62-d4f3-43af-88b2-26fbc725f50f-kube-api-access-tbltc\") pod \"ovn-operator-controller-manager-bbc5b68f9-d5d5r\" (UID: \"dacb8b62-d4f3-43af-88b2-26fbc725f50f\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.437224 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.449801 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn6zg\" (UniqueName: \"kubernetes.io/projected/a9a4fe04-37f1-47ad-ab32-e983f13df515-kube-api-access-hn6zg\") pod \"placement-operator-controller-manager-574d45c66c-m7zdz\" (UID: \"a9a4fe04-37f1-47ad-ab32-e983f13df515\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.450785 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fvrg\" (UniqueName: \"kubernetes.io/projected/256daf30-6602-4a5d-8a4b-0f96fb5450b2-kube-api-access-2fvrg\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.471144 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.487052 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn6zg\" (UniqueName: \"kubernetes.io/projected/a9a4fe04-37f1-47ad-ab32-e983f13df515-kube-api-access-hn6zg\") pod \"placement-operator-controller-manager-574d45c66c-m7zdz\" (UID: \"a9a4fe04-37f1-47ad-ab32-e983f13df515\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.491482 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.492833 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.492949 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.498100 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-flwnw" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.502461 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.503671 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.505734 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-txzxp" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.531148 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.532162 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.539120 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-st7cs" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.549248 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.551174 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-878d6\" (UniqueName: \"kubernetes.io/projected/a314a819-a02b-4e9f-a996-58bed1d2e6be-kube-api-access-878d6\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-jt5rp\" (UID: \"a314a819-a02b-4e9f-a996-58bed1d2e6be\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.551218 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bf6x\" (UniqueName: \"kubernetes.io/projected/e7f3a695-6649-482f-94d9-238966224ba8-kube-api-access-7bf6x\") pod \"swift-operator-controller-manager-677c674df7-hgmsc\" (UID: \"e7f3a695-6649-482f-94d9-238966224ba8\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.560568 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.560822 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.561590 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.562794 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert podName:722c5d70-ed42-4b75-8993-3b3643d44d1a nodeName:}" failed. No retries permitted until 2026-03-13 07:55:34.562776041 +0000 UTC m=+994.233555013 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert") pod "infra-operator-controller-manager-5995f4446f-lcwhv" (UID: "722c5d70-ed42-4b75-8993-3b3643d44d1a") : secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.565990 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.584492 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.585481 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.595618 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mtzkk" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.610592 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.653906 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.659445 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.662638 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfl7h\" (UniqueName: \"kubernetes.io/projected/78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e-kube-api-access-kfl7h\") pod \"test-operator-controller-manager-5c5cb9c4d7-jztw5\" (UID: \"78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.663026 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-878d6\" (UniqueName: \"kubernetes.io/projected/a314a819-a02b-4e9f-a996-58bed1d2e6be-kube-api-access-878d6\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-jt5rp\" (UID: \"a314a819-a02b-4e9f-a996-58bed1d2e6be\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.663140 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bf6x\" (UniqueName: \"kubernetes.io/projected/e7f3a695-6649-482f-94d9-238966224ba8-kube-api-access-7bf6x\") pod \"swift-operator-controller-manager-677c674df7-hgmsc\" (UID: \"e7f3a695-6649-482f-94d9-238966224ba8\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.663227 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqzxj\" (UniqueName: \"kubernetes.io/projected/3aefe2ee-3d59-46e9-9a6e-4f576566a49c-kube-api-access-fqzxj\") pod \"watcher-operator-controller-manager-6dd88c6f67-cm65k\" (UID: \"3aefe2ee-3d59-46e9-9a6e-4f576566a49c\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.671113 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-8n76x" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.671392 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.671503 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.683702 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.718669 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-878d6\" (UniqueName: \"kubernetes.io/projected/a314a819-a02b-4e9f-a996-58bed1d2e6be-kube-api-access-878d6\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-jt5rp\" (UID: \"a314a819-a02b-4e9f-a996-58bed1d2e6be\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.735294 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.736389 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.741360 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-kz5rn" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.751609 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bf6x\" (UniqueName: \"kubernetes.io/projected/e7f3a695-6649-482f-94d9-238966224ba8-kube-api-access-7bf6x\") pod \"swift-operator-controller-manager-677c674df7-hgmsc\" (UID: \"e7f3a695-6649-482f-94d9-238966224ba8\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.767397 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.767491 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.767550 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2k5z\" (UniqueName: \"kubernetes.io/projected/bad784ad-2cd8-4f50-832f-0e57d06249d6-kube-api-access-j2k5z\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.767615 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqzxj\" (UniqueName: \"kubernetes.io/projected/3aefe2ee-3d59-46e9-9a6e-4f576566a49c-kube-api-access-fqzxj\") pod \"watcher-operator-controller-manager-6dd88c6f67-cm65k\" (UID: \"3aefe2ee-3d59-46e9-9a6e-4f576566a49c\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.767689 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfl7h\" (UniqueName: \"kubernetes.io/projected/78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e-kube-api-access-kfl7h\") pod \"test-operator-controller-manager-5c5cb9c4d7-jztw5\" (UID: \"78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.774638 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq"] Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.786570 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.830612 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.850910 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.870096 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.870206 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2k5z\" (UniqueName: \"kubernetes.io/projected/bad784ad-2cd8-4f50-832f-0e57d06249d6-kube-api-access-j2k5z\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.870372 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhb9v\" (UniqueName: \"kubernetes.io/projected/8a6e1029-8d6f-4698-b92b-d0cd17bda62d-kube-api-access-nhb9v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tzshq\" (UID: \"8a6e1029-8d6f-4698-b92b-d0cd17bda62d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.870423 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.870437 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.870482 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.870511 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert podName:256daf30-6602-4a5d-8a4b-0f96fb5450b2 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:34.870489047 +0000 UTC m=+994.541268029 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" (UID: "256daf30-6602-4a5d-8a4b-0f96fb5450b2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.870639 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.870712 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:34.370687852 +0000 UTC m=+994.041466994 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "webhook-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.870896 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: E0313 07:55:33.870932 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:34.370922959 +0000 UTC m=+994.041701941 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "metrics-server-cert" not found Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.887414 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfl7h\" (UniqueName: \"kubernetes.io/projected/78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e-kube-api-access-kfl7h\") pod \"test-operator-controller-manager-5c5cb9c4d7-jztw5\" (UID: \"78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.924442 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqzxj\" (UniqueName: \"kubernetes.io/projected/3aefe2ee-3d59-46e9-9a6e-4f576566a49c-kube-api-access-fqzxj\") pod \"watcher-operator-controller-manager-6dd88c6f67-cm65k\" (UID: \"3aefe2ee-3d59-46e9-9a6e-4f576566a49c\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.947518 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.971396 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2k5z\" (UniqueName: \"kubernetes.io/projected/bad784ad-2cd8-4f50-832f-0e57d06249d6-kube-api-access-j2k5z\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:33 crc kubenswrapper[4876]: I0313 07:55:33.972397 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhb9v\" (UniqueName: \"kubernetes.io/projected/8a6e1029-8d6f-4698-b92b-d0cd17bda62d-kube-api-access-nhb9v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tzshq\" (UID: \"8a6e1029-8d6f-4698-b92b-d0cd17bda62d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.055740 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhb9v\" (UniqueName: \"kubernetes.io/projected/8a6e1029-8d6f-4698-b92b-d0cd17bda62d-kube-api-access-nhb9v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-tzshq\" (UID: \"8a6e1029-8d6f-4698-b92b-d0cd17bda62d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.059256 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2"] Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.182630 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" Mar 13 07:55:34 crc kubenswrapper[4876]: W0313 07:55:34.191294 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cfaf185_ed2d_4c36_9d38_fcf4e74905ec.slice/crio-c788b5a60596a32ff2be44e663d63b9a7bfff560dd76d0ce4863350c8f3c572a WatchSource:0}: Error finding container c788b5a60596a32ff2be44e663d63b9a7bfff560dd76d0ce4863350c8f3c572a: Status 404 returned error can't find the container with id c788b5a60596a32ff2be44e663d63b9a7bfff560dd76d0ce4863350c8f3c572a Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.302431 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" event={"ID":"4cfaf185-ed2d-4c36-9d38-fcf4e74905ec","Type":"ContainerStarted","Data":"c788b5a60596a32ff2be44e663d63b9a7bfff560dd76d0ce4863350c8f3c572a"} Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.308816 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.382918 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.382982 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:34 crc kubenswrapper[4876]: E0313 07:55:34.383186 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 07:55:34 crc kubenswrapper[4876]: E0313 07:55:34.383255 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:35.383223695 +0000 UTC m=+995.054002677 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "webhook-server-cert" not found Mar 13 07:55:34 crc kubenswrapper[4876]: E0313 07:55:34.383854 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 07:55:34 crc kubenswrapper[4876]: E0313 07:55:34.383938 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:35.383913435 +0000 UTC m=+995.054692487 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "metrics-server-cert" not found Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.468067 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7"] Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.585695 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:34 crc kubenswrapper[4876]: E0313 07:55:34.585935 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:34 crc kubenswrapper[4876]: E0313 07:55:34.585994 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert podName:722c5d70-ed42-4b75-8993-3b3643d44d1a nodeName:}" failed. No retries permitted until 2026-03-13 07:55:36.585976485 +0000 UTC m=+996.256755467 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert") pod "infra-operator-controller-manager-5995f4446f-lcwhv" (UID: "722c5d70-ed42-4b75-8993-3b3643d44d1a") : secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:34 crc kubenswrapper[4876]: W0313 07:55:34.661817 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d21a99d_11a5_4b5f_a44d_3f9cce84b82a.slice/crio-ebe3d48daea6a3c0baa86b7702b5d5de916835b78067cebd9a56ed7d96253829 WatchSource:0}: Error finding container ebe3d48daea6a3c0baa86b7702b5d5de916835b78067cebd9a56ed7d96253829: Status 404 returned error can't find the container with id ebe3d48daea6a3c0baa86b7702b5d5de916835b78067cebd9a56ed7d96253829 Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.791831 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd"] Mar 13 07:55:34 crc kubenswrapper[4876]: I0313 07:55:34.892877 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:34 crc kubenswrapper[4876]: E0313 07:55:34.893034 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:34 crc kubenswrapper[4876]: E0313 07:55:34.893088 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert podName:256daf30-6602-4a5d-8a4b-0f96fb5450b2 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:36.893072993 +0000 UTC m=+996.563851975 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" (UID: "256daf30-6602-4a5d-8a4b-0f96fb5450b2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.081577 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb"] Mar 13 07:55:35 crc kubenswrapper[4876]: W0313 07:55:35.092263 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ff4d1b1_aae8_4900_a317_164ea72a78b2.slice/crio-56f3d1d4e42f6ea115c49870e59a16c63f73c85ac407ddf419939aeb2b6ba682 WatchSource:0}: Error finding container 56f3d1d4e42f6ea115c49870e59a16c63f73c85ac407ddf419939aeb2b6ba682: Status 404 returned error can't find the container with id 56f3d1d4e42f6ea115c49870e59a16c63f73c85ac407ddf419939aeb2b6ba682 Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.329174 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" event={"ID":"8d21a99d-11a5-4b5f-a44d-3f9cce84b82a","Type":"ContainerStarted","Data":"ebe3d48daea6a3c0baa86b7702b5d5de916835b78067cebd9a56ed7d96253829"} Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.330470 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" event={"ID":"3ff4d1b1-aae8-4900-a317-164ea72a78b2","Type":"ContainerStarted","Data":"56f3d1d4e42f6ea115c49870e59a16c63f73c85ac407ddf419939aeb2b6ba682"} Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.334884 4876 generic.go:334] "Generic (PLEG): container finished" podID="74fa4506-343c-48a2-bb72-612cc2315daa" containerID="766ee45eb4c3f058b18cc7db21ce8286b744a277cb4ad505de38beb66edb1d72" exitCode=0 Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.334969 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6bj8" event={"ID":"74fa4506-343c-48a2-bb72-612cc2315daa","Type":"ContainerDied","Data":"766ee45eb4c3f058b18cc7db21ce8286b744a277cb4ad505de38beb66edb1d72"} Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.341847 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" event={"ID":"8c04b15f-3ca3-4875-9ea4-45244808af5f","Type":"ContainerStarted","Data":"c89e867c925e76d3af76d34f8370c5e9aa3e45b14322dc689110490176927d65"} Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.408723 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.409119 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.408918 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.409219 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:37.409198098 +0000 UTC m=+997.079977080 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "metrics-server-cert" not found Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.409279 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.409309 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:37.409299381 +0000 UTC m=+997.080078363 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "webhook-server-cert" not found Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.531181 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.556592 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.579495 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.594909 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.604486 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5"] Mar 13 07:55:35 crc kubenswrapper[4876]: W0313 07:55:35.618532 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a4fe04_37f1_47ad_ab32_e983f13df515.slice/crio-3743d1bba4c4fd753c5861c4f8cfa68354c4d2d94522324fc4f46b2661951854 WatchSource:0}: Error finding container 3743d1bba4c4fd753c5861c4f8cfa68354c4d2d94522324fc4f46b2661951854: Status 404 returned error can't find the container with id 3743d1bba4c4fd753c5861c4f8cfa68354c4d2d94522324fc4f46b2661951854 Mar 13 07:55:35 crc kubenswrapper[4876]: W0313 07:55:35.621667 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef934388_2da5_47fe_8679_ec520946de1b.slice/crio-af83f4fa1dc9508c0efec277bdf1656acc234a3334a40ebe7a67c4eea5f2fcad WatchSource:0}: Error finding container af83f4fa1dc9508c0efec277bdf1656acc234a3334a40ebe7a67c4eea5f2fcad: Status 404 returned error can't find the container with id af83f4fa1dc9508c0efec277bdf1656acc234a3334a40ebe7a67c4eea5f2fcad Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.627301 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.646605 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm"] Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.655084 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fqzxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6dd88c6f67-cm65k_openstack-operators(3aefe2ee-3d59-46e9-9a6e-4f576566a49c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.655103 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rwtgm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-684f77d66d-kmhm8_openstack-operators(33e3356e-590a-4b31-990f-4ecda38efb0a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.655669 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g9mql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-66d56f6ff4-897gd_openstack-operators(68223df4-94ca-47fe-afbc-93c77725ba2a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.655714 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz"] Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.656283 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" podUID="3aefe2ee-3d59-46e9-9a6e-4f576566a49c" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.658027 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" podUID="68223df4-94ca-47fe-afbc-93c77725ba2a" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.658085 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" podUID="33e3356e-590a-4b31-990f-4ecda38efb0a" Mar 13 07:55:35 crc kubenswrapper[4876]: W0313 07:55:35.658155 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a6e1029_8d6f_4698_b92b_d0cd17bda62d.slice/crio-2ec99d5b415f2da5c19c60e2c5a78f302f4ef3e11918b9f6d24a84b0ded0f162 WatchSource:0}: Error finding container 2ec99d5b415f2da5c19c60e2c5a78f302f4ef3e11918b9f6d24a84b0ded0f162: Status 404 returned error can't find the container with id 2ec99d5b415f2da5c19c60e2c5a78f302f4ef3e11918b9f6d24a84b0ded0f162 Mar 13 07:55:35 crc kubenswrapper[4876]: W0313 07:55:35.660103 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4284fe2d_d1bb_49f6_b555_0f46419ad11b.slice/crio-4a3d7b66ead7474eaa5a047b6b008930fdc364c03115d30870ba6cfb3db44f5d WatchSource:0}: Error finding container 4a3d7b66ead7474eaa5a047b6b008930fdc364c03115d30870ba6cfb3db44f5d: Status 404 returned error can't find the container with id 4a3d7b66ead7474eaa5a047b6b008930fdc364c03115d30870ba6cfb3db44f5d Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.662507 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh"] Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.663339 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vb8tt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-776c5696bf-lnb48_openstack-operators(4284fe2d-d1bb-49f6-b555-0f46419ad11b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.664508 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" podUID="4284fe2d-d1bb-49f6-b555-0f46419ad11b" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.666309 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nhb9v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-tzshq_openstack-operators(8a6e1029-8d6f-4698-b92b-d0cd17bda62d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.667467 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" podUID="8a6e1029-8d6f-4698-b92b-d0cd17bda62d" Mar 13 07:55:35 crc kubenswrapper[4876]: W0313 07:55:35.668344 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98ac6099_e731_43c8_8537_8095f1bfce09.slice/crio-a108dde4b4e6c1914358c8e1b29356cd956a38eaf67d7db6c1424c7d118d993d WatchSource:0}: Error finding container a108dde4b4e6c1914358c8e1b29356cd956a38eaf67d7db6c1424c7d118d993d: Status 404 returned error can't find the container with id a108dde4b4e6c1914358c8e1b29356cd956a38eaf67d7db6c1424c7d118d993d Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.668960 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8"] Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.669891 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5nq9r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-68f45f9d9f-bp9bt_openstack-operators(98ac6099-e731-43c8-8537-8095f1bfce09): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 07:55:35 crc kubenswrapper[4876]: E0313 07:55:35.671863 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" podUID="98ac6099-e731-43c8-8537-8095f1bfce09" Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.677480 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.684049 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.689143 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.699798 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.712461 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k"] Mar 13 07:55:35 crc kubenswrapper[4876]: I0313 07:55:35.717038 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq"] Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.359918 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" event={"ID":"68223df4-94ca-47fe-afbc-93c77725ba2a","Type":"ContainerStarted","Data":"4df9d32d2d18f68efc05cc5d0c2fd084e2161e8a45417fdf45a6c3f94a3cfb46"} Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.363341 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7\\\"\"" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" podUID="68223df4-94ca-47fe-afbc-93c77725ba2a" Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.370933 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6bj8" event={"ID":"74fa4506-343c-48a2-bb72-612cc2315daa","Type":"ContainerStarted","Data":"27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.380176 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" event={"ID":"3de3f00b-74de-4328-86e9-9f86c7277deb","Type":"ContainerStarted","Data":"87cabb90e2d3423b1c8600d33dd29159a126368fb77f9359bba04a427d93d7d0"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.397365 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" event={"ID":"33e3356e-590a-4b31-990f-4ecda38efb0a","Type":"ContainerStarted","Data":"fba04ce9140d8116d3f32df826c1a800967f470b59682f81affaa6332199fa79"} Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.401702 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" podUID="33e3356e-590a-4b31-990f-4ecda38efb0a" Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.402123 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" event={"ID":"dacb8b62-d4f3-43af-88b2-26fbc725f50f","Type":"ContainerStarted","Data":"680824b65937f19e7269027819fa4980c1ca1e1e3d43a08307b85c991a36911c"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.403806 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" event={"ID":"151ed757-e99e-4e7e-bfaa-7dec0ef8b038","Type":"ContainerStarted","Data":"c66161f2d056727cbeef8a66f28a8e84a9d41630d76b5aebd435997398da9a28"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.411724 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" event={"ID":"a314a819-a02b-4e9f-a996-58bed1d2e6be","Type":"ContainerStarted","Data":"61da3943fca1cc314448f515b434eb9d28f5ba7a46310d2843056a51bfff81a1"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.416387 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m6bj8" podStartSLOduration=2.950737876 podStartE2EDuration="5.41636485s" podCreationTimestamp="2026-03-13 07:55:31 +0000 UTC" firstStartedPulling="2026-03-13 07:55:33.348939169 +0000 UTC m=+993.019718151" lastFinishedPulling="2026-03-13 07:55:35.814566143 +0000 UTC m=+995.485345125" observedRunningTime="2026-03-13 07:55:36.413956192 +0000 UTC m=+996.084735194" watchObservedRunningTime="2026-03-13 07:55:36.41636485 +0000 UTC m=+996.087143832" Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.417026 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" event={"ID":"ef934388-2da5-47fe-8679-ec520946de1b","Type":"ContainerStarted","Data":"af83f4fa1dc9508c0efec277bdf1656acc234a3334a40ebe7a67c4eea5f2fcad"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.420431 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" event={"ID":"e7f3a695-6649-482f-94d9-238966224ba8","Type":"ContainerStarted","Data":"3732cd02beb91879723498597e691bb501c21aaac1740f1101646332743cc89f"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.423916 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" event={"ID":"6aeda87b-2ede-4dcb-9234-bfc8837f4ea8","Type":"ContainerStarted","Data":"738f5c980462dd5df128ac180050ba62aa02667c39a3cfd1f13fb0090b1cabd4"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.427103 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" event={"ID":"8a6e1029-8d6f-4698-b92b-d0cd17bda62d","Type":"ContainerStarted","Data":"2ec99d5b415f2da5c19c60e2c5a78f302f4ef3e11918b9f6d24a84b0ded0f162"} Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.428620 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" podUID="8a6e1029-8d6f-4698-b92b-d0cd17bda62d" Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.428773 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" event={"ID":"78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e","Type":"ContainerStarted","Data":"e5bec6f03f4ccde753af0b61043ad8483645e817bd51f9e4eabd62c9fcb9c940"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.430940 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" event={"ID":"98ac6099-e731-43c8-8537-8095f1bfce09","Type":"ContainerStarted","Data":"a108dde4b4e6c1914358c8e1b29356cd956a38eaf67d7db6c1424c7d118d993d"} Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.432790 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4\\\"\"" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" podUID="98ac6099-e731-43c8-8537-8095f1bfce09" Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.433208 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" event={"ID":"ff48d75e-19f2-4a48-bcd6-3855b494fdf1","Type":"ContainerStarted","Data":"90bb3997544be6e4be2ad61ef787c3c19211344bac3646dbdbbadf7ef43ea12d"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.435891 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" event={"ID":"a9a4fe04-37f1-47ad-ab32-e983f13df515","Type":"ContainerStarted","Data":"3743d1bba4c4fd753c5861c4f8cfa68354c4d2d94522324fc4f46b2661951854"} Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.447786 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" event={"ID":"4284fe2d-d1bb-49f6-b555-0f46419ad11b","Type":"ContainerStarted","Data":"4a3d7b66ead7474eaa5a047b6b008930fdc364c03115d30870ba6cfb3db44f5d"} Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.452421 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" podUID="4284fe2d-d1bb-49f6-b555-0f46419ad11b" Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.453442 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" event={"ID":"3aefe2ee-3d59-46e9-9a6e-4f576566a49c","Type":"ContainerStarted","Data":"abea4533ca66552f8c4a35860e24e772a7be5031e92de1365f164b0169bd327d"} Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.459541 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" podUID="3aefe2ee-3d59-46e9-9a6e-4f576566a49c" Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.644349 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.644600 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.644701 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert podName:722c5d70-ed42-4b75-8993-3b3643d44d1a nodeName:}" failed. No retries permitted until 2026-03-13 07:55:40.644676459 +0000 UTC m=+1000.315455451 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert") pod "infra-operator-controller-manager-5995f4446f-lcwhv" (UID: "722c5d70-ed42-4b75-8993-3b3643d44d1a") : secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:36 crc kubenswrapper[4876]: I0313 07:55:36.949113 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.949349 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:36 crc kubenswrapper[4876]: E0313 07:55:36.949473 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert podName:256daf30-6602-4a5d-8a4b-0f96fb5450b2 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:40.949446862 +0000 UTC m=+1000.620225854 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" (UID: "256daf30-6602-4a5d-8a4b-0f96fb5450b2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:37 crc kubenswrapper[4876]: I0313 07:55:37.460629 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:37 crc kubenswrapper[4876]: I0313 07:55:37.461066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.461263 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.461340 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:41.461319427 +0000 UTC m=+1001.132098409 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "metrics-server-cert" not found Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.461542 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.461605 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:41.461589234 +0000 UTC m=+1001.132368216 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "webhook-server-cert" not found Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.463492 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" podUID="8a6e1029-8d6f-4698-b92b-d0cd17bda62d" Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.463594 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7\\\"\"" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" podUID="68223df4-94ca-47fe-afbc-93c77725ba2a" Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.463751 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4\\\"\"" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" podUID="98ac6099-e731-43c8-8537-8095f1bfce09" Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.464307 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" podUID="4284fe2d-d1bb-49f6-b555-0f46419ad11b" Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.464399 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" podUID="3aefe2ee-3d59-46e9-9a6e-4f576566a49c" Mar 13 07:55:37 crc kubenswrapper[4876]: E0313 07:55:37.465572 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" podUID="33e3356e-590a-4b31-990f-4ecda38efb0a" Mar 13 07:55:40 crc kubenswrapper[4876]: I0313 07:55:40.657484 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:40 crc kubenswrapper[4876]: E0313 07:55:40.657794 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:40 crc kubenswrapper[4876]: E0313 07:55:40.658039 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert podName:722c5d70-ed42-4b75-8993-3b3643d44d1a nodeName:}" failed. No retries permitted until 2026-03-13 07:55:48.658014938 +0000 UTC m=+1008.328793920 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert") pod "infra-operator-controller-manager-5995f4446f-lcwhv" (UID: "722c5d70-ed42-4b75-8993-3b3643d44d1a") : secret "infra-operator-webhook-server-cert" not found Mar 13 07:55:40 crc kubenswrapper[4876]: I0313 07:55:40.962762 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:40 crc kubenswrapper[4876]: E0313 07:55:40.962987 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:40 crc kubenswrapper[4876]: E0313 07:55:40.963095 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert podName:256daf30-6602-4a5d-8a4b-0f96fb5450b2 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:48.963068699 +0000 UTC m=+1008.633847861 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" (UID: "256daf30-6602-4a5d-8a4b-0f96fb5450b2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.033085 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nfttj"] Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.040465 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.055252 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nfttj"] Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.164993 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn8q6\" (UniqueName: \"kubernetes.io/projected/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-kube-api-access-cn8q6\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.167041 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-catalog-content\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.167270 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-utilities\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.268685 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn8q6\" (UniqueName: \"kubernetes.io/projected/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-kube-api-access-cn8q6\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.268785 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-catalog-content\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.268834 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-utilities\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.269308 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-utilities\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.269554 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-catalog-content\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.297724 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn8q6\" (UniqueName: \"kubernetes.io/projected/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-kube-api-access-cn8q6\") pod \"community-operators-nfttj\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.369803 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.474109 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.474182 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:41 crc kubenswrapper[4876]: E0313 07:55:41.474333 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 07:55:41 crc kubenswrapper[4876]: E0313 07:55:41.474385 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:49.474369067 +0000 UTC m=+1009.145148049 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "webhook-server-cert" not found Mar 13 07:55:41 crc kubenswrapper[4876]: E0313 07:55:41.474475 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 07:55:41 crc kubenswrapper[4876]: E0313 07:55:41.474559 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:55:49.474537611 +0000 UTC m=+1009.145316764 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "metrics-server-cert" not found Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.944820 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:41 crc kubenswrapper[4876]: I0313 07:55:41.944890 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:42 crc kubenswrapper[4876]: I0313 07:55:42.003036 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:42 crc kubenswrapper[4876]: I0313 07:55:42.599925 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:55:44 crc kubenswrapper[4876]: I0313 07:55:44.400428 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6bj8"] Mar 13 07:55:45 crc kubenswrapper[4876]: I0313 07:55:45.538983 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m6bj8" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="registry-server" containerID="cri-o://27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4" gracePeriod=2 Mar 13 07:55:46 crc kubenswrapper[4876]: I0313 07:55:46.560382 4876 generic.go:334] "Generic (PLEG): container finished" podID="74fa4506-343c-48a2-bb72-612cc2315daa" containerID="27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4" exitCode=0 Mar 13 07:55:46 crc kubenswrapper[4876]: I0313 07:55:46.560439 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6bj8" event={"ID":"74fa4506-343c-48a2-bb72-612cc2315daa","Type":"ContainerDied","Data":"27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4"} Mar 13 07:55:48 crc kubenswrapper[4876]: I0313 07:55:48.715658 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:48 crc kubenswrapper[4876]: I0313 07:55:48.726123 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/722c5d70-ed42-4b75-8993-3b3643d44d1a-cert\") pod \"infra-operator-controller-manager-5995f4446f-lcwhv\" (UID: \"722c5d70-ed42-4b75-8993-3b3643d44d1a\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:49 crc kubenswrapper[4876]: I0313 07:55:49.012977 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:55:49 crc kubenswrapper[4876]: I0313 07:55:49.020983 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:49 crc kubenswrapper[4876]: I0313 07:55:49.026670 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/256daf30-6602-4a5d-8a4b-0f96fb5450b2-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q\" (UID: \"256daf30-6602-4a5d-8a4b-0f96fb5450b2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:49 crc kubenswrapper[4876]: I0313 07:55:49.306867 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:55:49 crc kubenswrapper[4876]: I0313 07:55:49.529066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:49 crc kubenswrapper[4876]: I0313 07:55:49.529517 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:55:49 crc kubenswrapper[4876]: E0313 07:55:49.529361 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 07:55:49 crc kubenswrapper[4876]: E0313 07:55:49.529759 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:56:05.529740592 +0000 UTC m=+1025.200519574 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "metrics-server-cert" not found Mar 13 07:55:49 crc kubenswrapper[4876]: E0313 07:55:49.529634 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 07:55:49 crc kubenswrapper[4876]: E0313 07:55:49.529893 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs podName:bad784ad-2cd8-4f50-832f-0e57d06249d6 nodeName:}" failed. No retries permitted until 2026-03-13 07:56:05.529884986 +0000 UTC m=+1025.200663968 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs") pod "openstack-operator-controller-manager-74f6d57fdc-mfkh6" (UID: "bad784ad-2cd8-4f50-832f-0e57d06249d6") : secret "webhook-server-cert" not found Mar 13 07:55:50 crc kubenswrapper[4876]: E0313 07:55:50.704482 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6" Mar 13 07:55:50 crc kubenswrapper[4876]: E0313 07:55:50.704731 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzb5q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6d9d6b584d-c6bfj_openstack-operators(ff48d75e-19f2-4a48-bcd6-3855b494fdf1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:55:50 crc kubenswrapper[4876]: E0313 07:55:50.705971 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" podUID="ff48d75e-19f2-4a48-bcd6-3855b494fdf1" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.404092 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nfcfw"] Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.409568 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.414418 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nfcfw"] Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.430832 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c" Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.431059 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7bf6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-677c674df7-hgmsc_openstack-operators(e7f3a695-6649-482f-94d9-238966224ba8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.432299 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" podUID="e7f3a695-6649-482f-94d9-238966224ba8" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.457912 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-catalog-content\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.457982 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-utilities\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.458008 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmdfk\" (UniqueName: \"kubernetes.io/projected/828564d0-2f02-42ba-a4b7-0ba9e06c987c-kube-api-access-zmdfk\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.558845 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-utilities\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.558899 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmdfk\" (UniqueName: \"kubernetes.io/projected/828564d0-2f02-42ba-a4b7-0ba9e06c987c-kube-api-access-zmdfk\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.559033 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-catalog-content\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.559663 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-catalog-content\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.560629 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-utilities\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.582538 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmdfk\" (UniqueName: \"kubernetes.io/projected/828564d0-2f02-42ba-a4b7-0ba9e06c987c-kube-api-access-zmdfk\") pod \"certified-operators-nfcfw\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.602227 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" podUID="ff48d75e-19f2-4a48-bcd6-3855b494fdf1" Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.603546 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" podUID="e7f3a695-6649-482f-94d9-238966224ba8" Mar 13 07:55:51 crc kubenswrapper[4876]: I0313 07:55:51.736040 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.946254 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4 is running failed: container process not found" containerID="27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.946773 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4 is running failed: container process not found" containerID="27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.947366 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4 is running failed: container process not found" containerID="27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 07:55:51 crc kubenswrapper[4876]: E0313 07:55:51.947410 4876 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-m6bj8" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="registry-server" Mar 13 07:55:53 crc kubenswrapper[4876]: E0313 07:55:53.015457 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:b99cd5e08bd85c6aaf717519187ba7bfeea359e1537d43b73a7364b7c38116e2" Mar 13 07:55:53 crc kubenswrapper[4876]: E0313 07:55:53.016032 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:b99cd5e08bd85c6aaf717519187ba7bfeea359e1537d43b73a7364b7c38116e2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-27v49,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-658d4cdd5-v7kpw_openstack-operators(6aeda87b-2ede-4dcb-9234-bfc8837f4ea8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:55:53 crc kubenswrapper[4876]: E0313 07:55:53.017301 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" podUID="6aeda87b-2ede-4dcb-9234-bfc8837f4ea8" Mar 13 07:55:53 crc kubenswrapper[4876]: E0313 07:55:53.625776 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:b99cd5e08bd85c6aaf717519187ba7bfeea359e1537d43b73a7364b7c38116e2\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" podUID="6aeda87b-2ede-4dcb-9234-bfc8837f4ea8" Mar 13 07:55:53 crc kubenswrapper[4876]: E0313 07:55:53.951899 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6" Mar 13 07:55:53 crc kubenswrapper[4876]: E0313 07:55:53.952135 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dfz4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-77b6666d85-jcjfd_openstack-operators(8c04b15f-3ca3-4875-9ea4-45244808af5f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:55:53 crc kubenswrapper[4876]: E0313 07:55:53.955291 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" podUID="8c04b15f-3ca3-4875-9ea4-45244808af5f" Mar 13 07:55:54 crc kubenswrapper[4876]: I0313 07:55:54.607522 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:55:54 crc kubenswrapper[4876]: I0313 07:55:54.607584 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:55:54 crc kubenswrapper[4876]: E0313 07:55:54.631980 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978" Mar 13 07:55:54 crc kubenswrapper[4876]: E0313 07:55:54.632201 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hn6zg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-574d45c66c-m7zdz_openstack-operators(a9a4fe04-37f1-47ad-ab32-e983f13df515): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:55:54 crc kubenswrapper[4876]: E0313 07:55:54.633533 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" podUID="a9a4fe04-37f1-47ad-ab32-e983f13df515" Mar 13 07:55:54 crc kubenswrapper[4876]: E0313 07:55:54.635714 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6\\\"\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" podUID="8c04b15f-3ca3-4875-9ea4-45244808af5f" Mar 13 07:55:55 crc kubenswrapper[4876]: E0313 07:55:55.557514 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" Mar 13 07:55:55 crc kubenswrapper[4876]: E0313 07:55:55.558031 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kfl7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-jztw5_openstack-operators(78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:55:55 crc kubenswrapper[4876]: E0313 07:55:55.559852 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" podUID="78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e" Mar 13 07:55:55 crc kubenswrapper[4876]: E0313 07:55:55.642882 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" podUID="a9a4fe04-37f1-47ad-ab32-e983f13df515" Mar 13 07:55:55 crc kubenswrapper[4876]: E0313 07:55:55.652771 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" podUID="78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e" Mar 13 07:55:59 crc kubenswrapper[4876]: E0313 07:55:59.680628 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922" Mar 13 07:55:59 crc kubenswrapper[4876]: E0313 07:55:59.681454 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x9k88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-569cc54c5-pfswh_openstack-operators(ef934388-2da5-47fe-8679-ec520946de1b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:55:59 crc kubenswrapper[4876]: E0313 07:55:59.683443 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" podUID="ef934388-2da5-47fe-8679-ec520946de1b" Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.153087 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556476-9x566"] Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.154175 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556476-9x566" Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.157192 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.157796 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.157934 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.166268 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556476-9x566"] Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.309959 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q75v2\" (UniqueName: \"kubernetes.io/projected/9431e26b-bb79-45a0-a0a2-9e11479cc64b-kube-api-access-q75v2\") pod \"auto-csr-approver-29556476-9x566\" (UID: \"9431e26b-bb79-45a0-a0a2-9e11479cc64b\") " pod="openshift-infra/auto-csr-approver-29556476-9x566" Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.411089 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q75v2\" (UniqueName: \"kubernetes.io/projected/9431e26b-bb79-45a0-a0a2-9e11479cc64b-kube-api-access-q75v2\") pod \"auto-csr-approver-29556476-9x566\" (UID: \"9431e26b-bb79-45a0-a0a2-9e11479cc64b\") " pod="openshift-infra/auto-csr-approver-29556476-9x566" Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.431884 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q75v2\" (UniqueName: \"kubernetes.io/projected/9431e26b-bb79-45a0-a0a2-9e11479cc64b-kube-api-access-q75v2\") pod \"auto-csr-approver-29556476-9x566\" (UID: \"9431e26b-bb79-45a0-a0a2-9e11479cc64b\") " pod="openshift-infra/auto-csr-approver-29556476-9x566" Mar 13 07:56:00 crc kubenswrapper[4876]: I0313 07:56:00.486716 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556476-9x566" Mar 13 07:56:00 crc kubenswrapper[4876]: E0313 07:56:00.680746 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922\\\"\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" podUID="ef934388-2da5-47fe-8679-ec520946de1b" Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.802652 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.830699 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-catalog-content\") pod \"74fa4506-343c-48a2-bb72-612cc2315daa\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.830758 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgfmt\" (UniqueName: \"kubernetes.io/projected/74fa4506-343c-48a2-bb72-612cc2315daa-kube-api-access-tgfmt\") pod \"74fa4506-343c-48a2-bb72-612cc2315daa\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.830804 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-utilities\") pod \"74fa4506-343c-48a2-bb72-612cc2315daa\" (UID: \"74fa4506-343c-48a2-bb72-612cc2315daa\") " Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.831914 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-utilities" (OuterVolumeSpecName: "utilities") pod "74fa4506-343c-48a2-bb72-612cc2315daa" (UID: "74fa4506-343c-48a2-bb72-612cc2315daa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.836913 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74fa4506-343c-48a2-bb72-612cc2315daa-kube-api-access-tgfmt" (OuterVolumeSpecName: "kube-api-access-tgfmt") pod "74fa4506-343c-48a2-bb72-612cc2315daa" (UID: "74fa4506-343c-48a2-bb72-612cc2315daa"). InnerVolumeSpecName "kube-api-access-tgfmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.860314 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74fa4506-343c-48a2-bb72-612cc2315daa" (UID: "74fa4506-343c-48a2-bb72-612cc2315daa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.932231 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.932294 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgfmt\" (UniqueName: \"kubernetes.io/projected/74fa4506-343c-48a2-bb72-612cc2315daa-kube-api-access-tgfmt\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:01 crc kubenswrapper[4876]: I0313 07:56:01.932311 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74fa4506-343c-48a2-bb72-612cc2315daa-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.729047 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nfttj"] Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.732132 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6bj8" event={"ID":"74fa4506-343c-48a2-bb72-612cc2315daa","Type":"ContainerDied","Data":"b2af2e3544a602c1d17a4e101613d26e8ba41b12441ff2d32d2f34bf72e6f574"} Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.732289 4876 scope.go:117] "RemoveContainer" containerID="27a5e4ff7e695697cf79b9a6bed47ccdbc7251b03c5aa93323f43d0850a28de4" Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.732509 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m6bj8" Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.737278 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" event={"ID":"3ff4d1b1-aae8-4900-a317-164ea72a78b2","Type":"ContainerStarted","Data":"aaa930d4ddf752b304b82e99c9291cfa685501b81bf7132292c665e6ab88c38c"} Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.738066 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.745819 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" event={"ID":"3de3f00b-74de-4328-86e9-9f86c7277deb","Type":"ContainerStarted","Data":"64d07030339b2a849843bcc085bbef60b6dd47e668697fa31739402b2a3a0db1"} Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.746277 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.765950 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" podStartSLOduration=8.399483657 podStartE2EDuration="30.76591934s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.095422522 +0000 UTC m=+994.766201504" lastFinishedPulling="2026-03-13 07:55:57.461858205 +0000 UTC m=+1017.132637187" observedRunningTime="2026-03-13 07:56:02.758432139 +0000 UTC m=+1022.429211121" watchObservedRunningTime="2026-03-13 07:56:02.76591934 +0000 UTC m=+1022.436698332" Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.790776 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556476-9x566"] Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.798658 4876 scope.go:117] "RemoveContainer" containerID="766ee45eb4c3f058b18cc7db21ce8286b744a277cb4ad505de38beb66edb1d72" Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.802718 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" podStartSLOduration=9.592438421 podStartE2EDuration="30.802677416s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.631095676 +0000 UTC m=+995.301874658" lastFinishedPulling="2026-03-13 07:55:56.841334671 +0000 UTC m=+1016.512113653" observedRunningTime="2026-03-13 07:56:02.77619705 +0000 UTC m=+1022.446976042" watchObservedRunningTime="2026-03-13 07:56:02.802677416 +0000 UTC m=+1022.473456398" Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.819872 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q"] Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.865717 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nfcfw"] Mar 13 07:56:02 crc kubenswrapper[4876]: I0313 07:56:02.881029 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv"] Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.087080 4876 scope.go:117] "RemoveContainer" containerID="6e548ddbf59d8c444facd3d827172806bf5e476b8006494b502cf04db3c1b714" Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.188404 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6bj8"] Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.202993 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6bj8"] Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.804024 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" event={"ID":"dacb8b62-d4f3-43af-88b2-26fbc725f50f","Type":"ContainerStarted","Data":"67d1f364b48ec880913497f49a491103064e33e0a3aae0c80dd0d54a350be069"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.804962 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.828494 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" event={"ID":"8a6e1029-8d6f-4698-b92b-d0cd17bda62d","Type":"ContainerStarted","Data":"958dd732c150fc26cbb4391a37a82af57e28a186a115c9e7e3b91a5742eda291"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.856798 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" event={"ID":"4284fe2d-d1bb-49f6-b555-0f46419ad11b","Type":"ContainerStarted","Data":"6fe32b805e374127a62b3375e9098bf57810299ce9edd87acce53909949381bc"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.857541 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.868585 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" podStartSLOduration=8.489311336 podStartE2EDuration="31.868569162s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.584865804 +0000 UTC m=+995.255644786" lastFinishedPulling="2026-03-13 07:55:58.96412363 +0000 UTC m=+1018.634902612" observedRunningTime="2026-03-13 07:56:03.86566925 +0000 UTC m=+1023.536448232" watchObservedRunningTime="2026-03-13 07:56:03.868569162 +0000 UTC m=+1023.539348144" Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.883936 4876 generic.go:334] "Generic (PLEG): container finished" podID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerID="a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05" exitCode=0 Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.884031 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfcfw" event={"ID":"828564d0-2f02-42ba-a4b7-0ba9e06c987c","Type":"ContainerDied","Data":"a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.884061 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfcfw" event={"ID":"828564d0-2f02-42ba-a4b7-0ba9e06c987c","Type":"ContainerStarted","Data":"14d1b34578a634659d778a1cf9273fa6c6bac8799138a0bc963bb25fa2cb303f"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.897216 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" event={"ID":"33e3356e-590a-4b31-990f-4ecda38efb0a","Type":"ContainerStarted","Data":"5075077df8a14b9ba86bbba2ff7f55ccb0d1656f0210f9e3a57e1aa1a2edfe69"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.897542 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.907675 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" event={"ID":"68223df4-94ca-47fe-afbc-93c77725ba2a","Type":"ContainerStarted","Data":"438f86e02a9753e1c2035901df70735a117bfb1024880940a1a89fef5cd57fd7"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.908982 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.919923 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-tzshq" podStartSLOduration=4.241541318 podStartE2EDuration="30.919903738s" podCreationTimestamp="2026-03-13 07:55:33 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.666209845 +0000 UTC m=+995.336988827" lastFinishedPulling="2026-03-13 07:56:02.344572265 +0000 UTC m=+1022.015351247" observedRunningTime="2026-03-13 07:56:03.918628302 +0000 UTC m=+1023.589407284" watchObservedRunningTime="2026-03-13 07:56:03.919903738 +0000 UTC m=+1023.590682720" Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.944543 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" event={"ID":"8d21a99d-11a5-4b5f-a44d-3f9cce84b82a","Type":"ContainerStarted","Data":"3d48b1e5f85aa9eb929731ef93e968455f19a82429ab337781fe173b840d1d4d"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.945329 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.960522 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" event={"ID":"4cfaf185-ed2d-4c36-9d38-fcf4e74905ec","Type":"ContainerStarted","Data":"bf27a654ef18c637736312c8cf36c0054b4b0ab59b061c0a5ee584553600b680"} Mar 13 07:56:03 crc kubenswrapper[4876]: I0313 07:56:03.961375 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:03.995888 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" podStartSLOduration=5.424520231 podStartE2EDuration="31.995870057s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.66317718 +0000 UTC m=+995.333956162" lastFinishedPulling="2026-03-13 07:56:02.234526986 +0000 UTC m=+1021.905305988" observedRunningTime="2026-03-13 07:56:03.972374905 +0000 UTC m=+1023.643153887" watchObservedRunningTime="2026-03-13 07:56:03.995870057 +0000 UTC m=+1023.666649039" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.006670 4876 generic.go:334] "Generic (PLEG): container finished" podID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerID="9a609f347f7a1a0b2285c394066f78c8680b2cc8eb8b76e920793d6d657d44cb" exitCode=0 Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.006739 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfttj" event={"ID":"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb","Type":"ContainerDied","Data":"9a609f347f7a1a0b2285c394066f78c8680b2cc8eb8b76e920793d6d657d44cb"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.006895 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfttj" event={"ID":"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb","Type":"ContainerStarted","Data":"16ba82d3acec5ea18482cbd4fafc33de2f426c1250d85d508dbc94ad24bb1231"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.024850 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" event={"ID":"151ed757-e99e-4e7e-bfaa-7dec0ef8b038","Type":"ContainerStarted","Data":"eee5d97b0377ce041c96c2b67c72d8b04f98f689bae85d275635d08b6134a0c8"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.025857 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.046954 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" podStartSLOduration=9.248869687 podStartE2EDuration="32.041231284s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:34.669500278 +0000 UTC m=+994.340279260" lastFinishedPulling="2026-03-13 07:55:57.461861875 +0000 UTC m=+1017.132640857" observedRunningTime="2026-03-13 07:56:04.028847785 +0000 UTC m=+1023.699626767" watchObservedRunningTime="2026-03-13 07:56:04.041231284 +0000 UTC m=+1023.712010266" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.064041 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" event={"ID":"a314a819-a02b-4e9f-a996-58bed1d2e6be","Type":"ContainerStarted","Data":"66d9b703c8ebc57663ac97e0809c668c2c5f59e9b0837878fab2a6f0ce8605b1"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.068314 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.068549 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" event={"ID":"3aefe2ee-3d59-46e9-9a6e-4f576566a49c","Type":"ContainerStarted","Data":"b7e7ab85177622d41c338daef2c43415ac39af9f2bf1ff7d4d0b881f89c56bf6"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.069284 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.070886 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" event={"ID":"98ac6099-e731-43c8-8537-8095f1bfce09","Type":"ContainerStarted","Data":"a24207b1f3f689097d2126e0a9e7cf00661f8324a47cf79d0a6df1a2b2d9553f"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.071978 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.073020 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556476-9x566" event={"ID":"9431e26b-bb79-45a0-a0a2-9e11479cc64b","Type":"ContainerStarted","Data":"7e6f7868df8543edf9b1b8681fea7e25018f309468adb906f6758803b11d9634"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.084147 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" podStartSLOduration=5.452533658 podStartE2EDuration="32.084127192s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.654971388 +0000 UTC m=+995.325750360" lastFinishedPulling="2026-03-13 07:56:02.286564912 +0000 UTC m=+1021.957343894" observedRunningTime="2026-03-13 07:56:04.063738938 +0000 UTC m=+1023.734517920" watchObservedRunningTime="2026-03-13 07:56:04.084127192 +0000 UTC m=+1023.754906174" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.099572 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" event={"ID":"722c5d70-ed42-4b75-8993-3b3643d44d1a","Type":"ContainerStarted","Data":"ab6cd9b9b81179339311b54ee7a4025fb7d029098e1062f65e039ca92fdd359e"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.116463 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" podStartSLOduration=5.455068431 podStartE2EDuration="32.116440022s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.655147744 +0000 UTC m=+995.325926726" lastFinishedPulling="2026-03-13 07:56:02.316519335 +0000 UTC m=+1021.987298317" observedRunningTime="2026-03-13 07:56:04.094809102 +0000 UTC m=+1023.765588084" watchObservedRunningTime="2026-03-13 07:56:04.116440022 +0000 UTC m=+1023.787219004" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.134047 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" event={"ID":"256daf30-6602-4a5d-8a4b-0f96fb5450b2","Type":"ContainerStarted","Data":"38ffbd6fda2aa489eebd8eb55345791b00e516d3ec2d9e6bd44d7d61badd992a"} Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.172006 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" podStartSLOduration=9.567168179 podStartE2EDuration="32.171971246s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:34.236710069 +0000 UTC m=+993.907489051" lastFinishedPulling="2026-03-13 07:55:56.841513146 +0000 UTC m=+1016.512292118" observedRunningTime="2026-03-13 07:56:04.170723321 +0000 UTC m=+1023.841502293" watchObservedRunningTime="2026-03-13 07:56:04.171971246 +0000 UTC m=+1023.842750228" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.327748 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" podStartSLOduration=7.998080725 podStartE2EDuration="31.327718162s" podCreationTimestamp="2026-03-13 07:55:33 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.654358342 +0000 UTC m=+995.325137324" lastFinishedPulling="2026-03-13 07:55:58.983995769 +0000 UTC m=+1018.654774761" observedRunningTime="2026-03-13 07:56:04.321828836 +0000 UTC m=+1023.992607828" watchObservedRunningTime="2026-03-13 07:56:04.327718162 +0000 UTC m=+1023.998497154" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.335594 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" podStartSLOduration=5.762703913 podStartE2EDuration="32.335565782s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.669751305 +0000 UTC m=+995.340530287" lastFinishedPulling="2026-03-13 07:56:02.242613174 +0000 UTC m=+1021.913392156" observedRunningTime="2026-03-13 07:56:04.252154984 +0000 UTC m=+1023.922933966" watchObservedRunningTime="2026-03-13 07:56:04.335565782 +0000 UTC m=+1024.006344764" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.368718 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" podStartSLOduration=11.10089506 podStartE2EDuration="32.368693396s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.573679789 +0000 UTC m=+995.244458761" lastFinishedPulling="2026-03-13 07:55:56.841478115 +0000 UTC m=+1016.512257097" observedRunningTime="2026-03-13 07:56:04.354124496 +0000 UTC m=+1024.024903478" watchObservedRunningTime="2026-03-13 07:56:04.368693396 +0000 UTC m=+1024.039472378" Mar 13 07:56:04 crc kubenswrapper[4876]: I0313 07:56:04.381213 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" podStartSLOduration=4.801462036 podStartE2EDuration="31.381196548s" podCreationTimestamp="2026-03-13 07:55:33 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.654946948 +0000 UTC m=+995.325725930" lastFinishedPulling="2026-03-13 07:56:02.23468147 +0000 UTC m=+1021.905460442" observedRunningTime="2026-03-13 07:56:04.37880404 +0000 UTC m=+1024.049583022" watchObservedRunningTime="2026-03-13 07:56:04.381196548 +0000 UTC m=+1024.051975520" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.046521 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" path="/var/lib/kubelet/pods/74fa4506-343c-48a2-bb72-612cc2315daa/volumes" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.176387 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556476-9x566" event={"ID":"9431e26b-bb79-45a0-a0a2-9e11479cc64b","Type":"ContainerStarted","Data":"ae6a3a4dc748772291167dd4abc6cbf51ceb60eb996b6bb924f1925ce25dc20f"} Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.179309 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" event={"ID":"ff48d75e-19f2-4a48-bcd6-3855b494fdf1","Type":"ContainerStarted","Data":"ea9a69b6a22b7aa5894337c3e197a57ef9178669ec136ecfc6f428d5afe8e43c"} Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.179616 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.187393 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfcfw" event={"ID":"828564d0-2f02-42ba-a4b7-0ba9e06c987c","Type":"ContainerStarted","Data":"ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58"} Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.194466 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfttj" event={"ID":"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb","Type":"ContainerStarted","Data":"d970a7e3d7684e707734ce94a0a49a450bf8b211ad600766eb2c36d2c0f36608"} Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.215797 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556476-9x566" podStartSLOduration=3.839042561 podStartE2EDuration="5.21577536s" podCreationTimestamp="2026-03-13 07:56:00 +0000 UTC" firstStartedPulling="2026-03-13 07:56:02.881741732 +0000 UTC m=+1022.552520714" lastFinishedPulling="2026-03-13 07:56:04.258474541 +0000 UTC m=+1023.929253513" observedRunningTime="2026-03-13 07:56:05.199673807 +0000 UTC m=+1024.870452789" watchObservedRunningTime="2026-03-13 07:56:05.21577536 +0000 UTC m=+1024.886554342" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.265744 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" podStartSLOduration=4.192397773 podStartE2EDuration="33.265717847s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.60888342 +0000 UTC m=+995.279662402" lastFinishedPulling="2026-03-13 07:56:04.682203494 +0000 UTC m=+1024.352982476" observedRunningTime="2026-03-13 07:56:05.263704439 +0000 UTC m=+1024.934483421" watchObservedRunningTime="2026-03-13 07:56:05.265717847 +0000 UTC m=+1024.936496829" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.540039 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.540121 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.552363 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-metrics-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.565571 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bad784ad-2cd8-4f50-832f-0e57d06249d6-webhook-certs\") pod \"openstack-operator-controller-manager-74f6d57fdc-mfkh6\" (UID: \"bad784ad-2cd8-4f50-832f-0e57d06249d6\") " pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.835888 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-8n76x" Mar 13 07:56:05 crc kubenswrapper[4876]: I0313 07:56:05.844953 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:56:06 crc kubenswrapper[4876]: I0313 07:56:06.212484 4876 generic.go:334] "Generic (PLEG): container finished" podID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerID="d970a7e3d7684e707734ce94a0a49a450bf8b211ad600766eb2c36d2c0f36608" exitCode=0 Mar 13 07:56:06 crc kubenswrapper[4876]: I0313 07:56:06.212561 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfttj" event={"ID":"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb","Type":"ContainerDied","Data":"d970a7e3d7684e707734ce94a0a49a450bf8b211ad600766eb2c36d2c0f36608"} Mar 13 07:56:06 crc kubenswrapper[4876]: I0313 07:56:06.219959 4876 generic.go:334] "Generic (PLEG): container finished" podID="9431e26b-bb79-45a0-a0a2-9e11479cc64b" containerID="ae6a3a4dc748772291167dd4abc6cbf51ceb60eb996b6bb924f1925ce25dc20f" exitCode=0 Mar 13 07:56:06 crc kubenswrapper[4876]: I0313 07:56:06.220064 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556476-9x566" event={"ID":"9431e26b-bb79-45a0-a0a2-9e11479cc64b","Type":"ContainerDied","Data":"ae6a3a4dc748772291167dd4abc6cbf51ceb60eb996b6bb924f1925ce25dc20f"} Mar 13 07:56:06 crc kubenswrapper[4876]: I0313 07:56:06.223458 4876 generic.go:334] "Generic (PLEG): container finished" podID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerID="ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58" exitCode=0 Mar 13 07:56:06 crc kubenswrapper[4876]: I0313 07:56:06.225170 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfcfw" event={"ID":"828564d0-2f02-42ba-a4b7-0ba9e06c987c","Type":"ContainerDied","Data":"ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58"} Mar 13 07:56:06 crc kubenswrapper[4876]: I0313 07:56:06.271200 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6"] Mar 13 07:56:07 crc kubenswrapper[4876]: I0313 07:56:07.255885 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" event={"ID":"bad784ad-2cd8-4f50-832f-0e57d06249d6","Type":"ContainerStarted","Data":"410f241d7fd8f698bbd1bc5bf1708a5055f74a1f6d0b7673d6c861f7edff4ddc"} Mar 13 07:56:07 crc kubenswrapper[4876]: I0313 07:56:07.256306 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:56:07 crc kubenswrapper[4876]: I0313 07:56:07.256319 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" event={"ID":"bad784ad-2cd8-4f50-832f-0e57d06249d6","Type":"ContainerStarted","Data":"d6eee2a5e40c11040b84a37a70f8f6936baee003c4ad54951faab0120ec46ae2"} Mar 13 07:56:07 crc kubenswrapper[4876]: I0313 07:56:07.290770 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" podStartSLOduration=34.290744122 podStartE2EDuration="34.290744122s" podCreationTimestamp="2026-03-13 07:55:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:56:07.284816315 +0000 UTC m=+1026.955595297" watchObservedRunningTime="2026-03-13 07:56:07.290744122 +0000 UTC m=+1026.961523104" Mar 13 07:56:08 crc kubenswrapper[4876]: I0313 07:56:08.265492 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556476-9x566" event={"ID":"9431e26b-bb79-45a0-a0a2-9e11479cc64b","Type":"ContainerDied","Data":"7e6f7868df8543edf9b1b8681fea7e25018f309468adb906f6758803b11d9634"} Mar 13 07:56:08 crc kubenswrapper[4876]: I0313 07:56:08.265866 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e6f7868df8543edf9b1b8681fea7e25018f309468adb906f6758803b11d9634" Mar 13 07:56:08 crc kubenswrapper[4876]: I0313 07:56:08.275311 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556476-9x566" Mar 13 07:56:08 crc kubenswrapper[4876]: I0313 07:56:08.429460 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q75v2\" (UniqueName: \"kubernetes.io/projected/9431e26b-bb79-45a0-a0a2-9e11479cc64b-kube-api-access-q75v2\") pod \"9431e26b-bb79-45a0-a0a2-9e11479cc64b\" (UID: \"9431e26b-bb79-45a0-a0a2-9e11479cc64b\") " Mar 13 07:56:08 crc kubenswrapper[4876]: I0313 07:56:08.437734 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9431e26b-bb79-45a0-a0a2-9e11479cc64b-kube-api-access-q75v2" (OuterVolumeSpecName: "kube-api-access-q75v2") pod "9431e26b-bb79-45a0-a0a2-9e11479cc64b" (UID: "9431e26b-bb79-45a0-a0a2-9e11479cc64b"). InnerVolumeSpecName "kube-api-access-q75v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:56:08 crc kubenswrapper[4876]: I0313 07:56:08.531344 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q75v2\" (UniqueName: \"kubernetes.io/projected/9431e26b-bb79-45a0-a0a2-9e11479cc64b-kube-api-access-q75v2\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.276388 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfttj" event={"ID":"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb","Type":"ContainerStarted","Data":"3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a"} Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.277726 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" event={"ID":"722c5d70-ed42-4b75-8993-3b3643d44d1a","Type":"ContainerStarted","Data":"59e7e2252190d1b8281dab217a22346132406b4efd0ce06175a8180e7418fe44"} Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.278447 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.279470 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" event={"ID":"256daf30-6602-4a5d-8a4b-0f96fb5450b2","Type":"ContainerStarted","Data":"2e5ac0cc9fb4985cdf35dde6be4e996628aff879f7bf229e51881b7595a90c8a"} Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.279916 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.282063 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" event={"ID":"e7f3a695-6649-482f-94d9-238966224ba8","Type":"ContainerStarted","Data":"ffc8e7e0ab94e69b91bd9c539302971c07d5687cfc07ca46343b3e5c57cd6ee6"} Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.282400 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.283906 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfcfw" event={"ID":"828564d0-2f02-42ba-a4b7-0ba9e06c987c","Type":"ContainerStarted","Data":"315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5"} Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.285220 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" event={"ID":"6aeda87b-2ede-4dcb-9234-bfc8837f4ea8","Type":"ContainerStarted","Data":"ac6c48c9401298385c1783f7de8ce4ee82df8be5c97de9a83468c0ac36673281"} Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.285412 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.286528 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556476-9x566" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.286938 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" event={"ID":"78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e","Type":"ContainerStarted","Data":"797857cac6bc7163588012fb0139a4185b174124dbbdbf4eac397fd49921aaf2"} Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.287360 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.303921 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nfttj" podStartSLOduration=23.901233223 podStartE2EDuration="28.303892034s" podCreationTimestamp="2026-03-13 07:55:41 +0000 UTC" firstStartedPulling="2026-03-13 07:56:04.012277529 +0000 UTC m=+1023.683056511" lastFinishedPulling="2026-03-13 07:56:08.41493634 +0000 UTC m=+1028.085715322" observedRunningTime="2026-03-13 07:56:09.298494612 +0000 UTC m=+1028.969273594" watchObservedRunningTime="2026-03-13 07:56:09.303892034 +0000 UTC m=+1028.974671026" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.339530 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" podStartSLOduration=5.517616212 podStartE2EDuration="36.339511556s" podCreationTimestamp="2026-03-13 07:55:33 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.654383382 +0000 UTC m=+995.325162364" lastFinishedPulling="2026-03-13 07:56:06.476278726 +0000 UTC m=+1026.147057708" observedRunningTime="2026-03-13 07:56:09.336731048 +0000 UTC m=+1029.007510030" watchObservedRunningTime="2026-03-13 07:56:09.339511556 +0000 UTC m=+1029.010290538" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.359078 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556470-k8djl"] Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.364637 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556470-k8djl"] Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.368684 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" podStartSLOduration=3.594213508 podStartE2EDuration="36.368658677s" podCreationTimestamp="2026-03-13 07:55:33 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.605445474 +0000 UTC m=+995.276224456" lastFinishedPulling="2026-03-13 07:56:08.379890643 +0000 UTC m=+1028.050669625" observedRunningTime="2026-03-13 07:56:09.359743477 +0000 UTC m=+1029.030522459" watchObservedRunningTime="2026-03-13 07:56:09.368658677 +0000 UTC m=+1029.039437659" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.393210 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" podStartSLOduration=31.915465062 podStartE2EDuration="37.393189509s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:56:02.885003883 +0000 UTC m=+1022.555782875" lastFinishedPulling="2026-03-13 07:56:08.36272834 +0000 UTC m=+1028.033507322" observedRunningTime="2026-03-13 07:56:09.388471965 +0000 UTC m=+1029.059250937" watchObservedRunningTime="2026-03-13 07:56:09.393189509 +0000 UTC m=+1029.063968501" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.421651 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nfcfw" podStartSLOduration=13.932495933 podStartE2EDuration="18.421634619s" podCreationTimestamp="2026-03-13 07:55:51 +0000 UTC" firstStartedPulling="2026-03-13 07:56:03.893660829 +0000 UTC m=+1023.564439811" lastFinishedPulling="2026-03-13 07:56:08.382799515 +0000 UTC m=+1028.053578497" observedRunningTime="2026-03-13 07:56:09.418285176 +0000 UTC m=+1029.089064158" watchObservedRunningTime="2026-03-13 07:56:09.421634619 +0000 UTC m=+1029.092413601" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.443506 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" podStartSLOduration=6.5698120939999995 podStartE2EDuration="37.443475635s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.595297959 +0000 UTC m=+995.266076941" lastFinishedPulling="2026-03-13 07:56:06.4689615 +0000 UTC m=+1026.139740482" observedRunningTime="2026-03-13 07:56:09.435656564 +0000 UTC m=+1029.106435546" watchObservedRunningTime="2026-03-13 07:56:09.443475635 +0000 UTC m=+1029.114254617" Mar 13 07:56:09 crc kubenswrapper[4876]: I0313 07:56:09.462689 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" podStartSLOduration=32.005721745 podStartE2EDuration="37.462664605s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:56:02.923166309 +0000 UTC m=+1022.593945291" lastFinishedPulling="2026-03-13 07:56:08.380109169 +0000 UTC m=+1028.050888151" observedRunningTime="2026-03-13 07:56:09.458181169 +0000 UTC m=+1029.128960151" watchObservedRunningTime="2026-03-13 07:56:09.462664605 +0000 UTC m=+1029.133443587" Mar 13 07:56:10 crc kubenswrapper[4876]: I0313 07:56:10.295197 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" event={"ID":"8c04b15f-3ca3-4875-9ea4-45244808af5f","Type":"ContainerStarted","Data":"bc815a75a6b0054db99bbfd152e9d983e4e1c39d092208739ea60cc470858275"} Mar 13 07:56:10 crc kubenswrapper[4876]: I0313 07:56:10.295902 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" Mar 13 07:56:10 crc kubenswrapper[4876]: I0313 07:56:10.296874 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" event={"ID":"a9a4fe04-37f1-47ad-ab32-e983f13df515","Type":"ContainerStarted","Data":"aaafa9bdeaf96fca434d41213d2e176faa2f81e187f77eed6767c7a33270c1c3"} Mar 13 07:56:10 crc kubenswrapper[4876]: I0313 07:56:10.322926 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" podStartSLOduration=3.696236621 podStartE2EDuration="38.3229084s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:34.943357479 +0000 UTC m=+994.614136461" lastFinishedPulling="2026-03-13 07:56:09.570029258 +0000 UTC m=+1029.240808240" observedRunningTime="2026-03-13 07:56:10.320467411 +0000 UTC m=+1029.991246393" watchObservedRunningTime="2026-03-13 07:56:10.3229084 +0000 UTC m=+1029.993687382" Mar 13 07:56:11 crc kubenswrapper[4876]: I0313 07:56:11.044497 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34633e40-856c-4f46-a981-1cb41f46dacd" path="/var/lib/kubelet/pods/34633e40-856c-4f46-a981-1cb41f46dacd/volumes" Mar 13 07:56:11 crc kubenswrapper[4876]: I0313 07:56:11.371027 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:56:11 crc kubenswrapper[4876]: I0313 07:56:11.371348 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:56:11 crc kubenswrapper[4876]: I0313 07:56:11.438001 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:56:11 crc kubenswrapper[4876]: I0313 07:56:11.461768 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" podStartSLOduration=4.521392588 podStartE2EDuration="38.46174764s" podCreationTimestamp="2026-03-13 07:55:33 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.631274271 +0000 UTC m=+995.302053253" lastFinishedPulling="2026-03-13 07:56:09.571629323 +0000 UTC m=+1029.242408305" observedRunningTime="2026-03-13 07:56:10.34812485 +0000 UTC m=+1030.018903832" watchObservedRunningTime="2026-03-13 07:56:11.46174764 +0000 UTC m=+1031.132526622" Mar 13 07:56:11 crc kubenswrapper[4876]: I0313 07:56:11.736577 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:56:11 crc kubenswrapper[4876]: I0313 07:56:11.736624 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:56:11 crc kubenswrapper[4876]: I0313 07:56:11.781599 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:56:12 crc kubenswrapper[4876]: I0313 07:56:12.946102 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-6slg2" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.059875 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-897gd" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.059989 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w45l7" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.080179 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-zlt8c" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.160526 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-c6bfj" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.225128 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-n5stb" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.230863 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-bp9bt" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.259039 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-kmhm8" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.279460 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-v7kpw" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.363405 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.442019 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-lnb48" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.474771 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-d5d5r" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.565035 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-m2xqm" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.788381 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.834410 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-677c674df7-hgmsc" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.861165 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-jt5rp" Mar 13 07:56:13 crc kubenswrapper[4876]: I0313 07:56:13.950972 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cm65k" Mar 13 07:56:14 crc kubenswrapper[4876]: I0313 07:56:14.004364 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nfttj"] Mar 13 07:56:14 crc kubenswrapper[4876]: I0313 07:56:14.186463 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jztw5" Mar 13 07:56:15 crc kubenswrapper[4876]: I0313 07:56:15.327846 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nfttj" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="registry-server" containerID="cri-o://3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a" gracePeriod=2 Mar 13 07:56:15 crc kubenswrapper[4876]: I0313 07:56:15.857375 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-74f6d57fdc-mfkh6" Mar 13 07:56:17 crc kubenswrapper[4876]: I0313 07:56:17.352838 4876 generic.go:334] "Generic (PLEG): container finished" podID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerID="3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a" exitCode=0 Mar 13 07:56:17 crc kubenswrapper[4876]: I0313 07:56:17.352949 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfttj" event={"ID":"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb","Type":"ContainerDied","Data":"3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a"} Mar 13 07:56:19 crc kubenswrapper[4876]: I0313 07:56:19.024990 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-lcwhv" Mar 13 07:56:19 crc kubenswrapper[4876]: I0313 07:56:19.315864 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q" Mar 13 07:56:21 crc kubenswrapper[4876]: E0313 07:56:21.372444 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a is running failed: container process not found" containerID="3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 07:56:21 crc kubenswrapper[4876]: E0313 07:56:21.372979 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a is running failed: container process not found" containerID="3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 07:56:21 crc kubenswrapper[4876]: E0313 07:56:21.373334 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a is running failed: container process not found" containerID="3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 07:56:21 crc kubenswrapper[4876]: E0313 07:56:21.373360 4876 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-nfttj" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="registry-server" Mar 13 07:56:21 crc kubenswrapper[4876]: I0313 07:56:21.795953 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:56:21 crc kubenswrapper[4876]: I0313 07:56:21.853812 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nfcfw"] Mar 13 07:56:22 crc kubenswrapper[4876]: I0313 07:56:22.426031 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nfcfw" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerName="registry-server" containerID="cri-o://315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5" gracePeriod=2 Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.111848 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-jcjfd" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.269326 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.332615 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn8q6\" (UniqueName: \"kubernetes.io/projected/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-kube-api-access-cn8q6\") pod \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.332763 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-catalog-content\") pod \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.332915 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-utilities\") pod \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\" (UID: \"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb\") " Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.335366 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-utilities" (OuterVolumeSpecName: "utilities") pod "6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" (UID: "6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.353675 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-kube-api-access-cn8q6" (OuterVolumeSpecName: "kube-api-access-cn8q6") pod "6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" (UID: "6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb"). InnerVolumeSpecName "kube-api-access-cn8q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.399187 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" (UID: "6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.417593 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.434554 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-catalog-content\") pod \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.434616 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmdfk\" (UniqueName: \"kubernetes.io/projected/828564d0-2f02-42ba-a4b7-0ba9e06c987c-kube-api-access-zmdfk\") pod \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.434695 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-utilities\") pod \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\" (UID: \"828564d0-2f02-42ba-a4b7-0ba9e06c987c\") " Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.435395 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.435417 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn8q6\" (UniqueName: \"kubernetes.io/projected/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-kube-api-access-cn8q6\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.435431 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.436021 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-utilities" (OuterVolumeSpecName: "utilities") pod "828564d0-2f02-42ba-a4b7-0ba9e06c987c" (UID: "828564d0-2f02-42ba-a4b7-0ba9e06c987c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.439474 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/828564d0-2f02-42ba-a4b7-0ba9e06c987c-kube-api-access-zmdfk" (OuterVolumeSpecName: "kube-api-access-zmdfk") pod "828564d0-2f02-42ba-a4b7-0ba9e06c987c" (UID: "828564d0-2f02-42ba-a4b7-0ba9e06c987c"). InnerVolumeSpecName "kube-api-access-zmdfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.448137 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" event={"ID":"ef934388-2da5-47fe-8679-ec520946de1b","Type":"ContainerStarted","Data":"619caf1879226eb22a7ba08fde40ec9221a280a6ec4964589586fea99c5b6663"} Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.448514 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.460131 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfttj" event={"ID":"6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb","Type":"ContainerDied","Data":"16ba82d3acec5ea18482cbd4fafc33de2f426c1250d85d508dbc94ad24bb1231"} Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.460200 4876 scope.go:117] "RemoveContainer" containerID="3a2f4bb6d3ad23db8253740a0835b5c181e73aa48244460e612c188110a1c12a" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.460368 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nfttj" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.466987 4876 generic.go:334] "Generic (PLEG): container finished" podID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerID="315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5" exitCode=0 Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.467053 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfcfw" event={"ID":"828564d0-2f02-42ba-a4b7-0ba9e06c987c","Type":"ContainerDied","Data":"315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5"} Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.467101 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfcfw" event={"ID":"828564d0-2f02-42ba-a4b7-0ba9e06c987c","Type":"ContainerDied","Data":"14d1b34578a634659d778a1cf9273fa6c6bac8799138a0bc963bb25fa2cb303f"} Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.467198 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nfcfw" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.473323 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" podStartSLOduration=4.111953667 podStartE2EDuration="51.473296344s" podCreationTimestamp="2026-03-13 07:55:32 +0000 UTC" firstStartedPulling="2026-03-13 07:55:35.645603025 +0000 UTC m=+995.316381997" lastFinishedPulling="2026-03-13 07:56:23.006945672 +0000 UTC m=+1042.677724674" observedRunningTime="2026-03-13 07:56:23.465937157 +0000 UTC m=+1043.136716139" watchObservedRunningTime="2026-03-13 07:56:23.473296344 +0000 UTC m=+1043.144075346" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.495435 4876 scope.go:117] "RemoveContainer" containerID="d970a7e3d7684e707734ce94a0a49a450bf8b211ad600766eb2c36d2c0f36608" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.501226 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nfttj"] Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.506385 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nfttj"] Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.516338 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "828564d0-2f02-42ba-a4b7-0ba9e06c987c" (UID: "828564d0-2f02-42ba-a4b7-0ba9e06c987c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.525954 4876 scope.go:117] "RemoveContainer" containerID="9a609f347f7a1a0b2285c394066f78c8680b2cc8eb8b76e920793d6d657d44cb" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.535955 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.536066 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmdfk\" (UniqueName: \"kubernetes.io/projected/828564d0-2f02-42ba-a4b7-0ba9e06c987c-kube-api-access-zmdfk\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.536144 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828564d0-2f02-42ba-a4b7-0ba9e06c987c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.547649 4876 scope.go:117] "RemoveContainer" containerID="315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.568129 4876 scope.go:117] "RemoveContainer" containerID="ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.589911 4876 scope.go:117] "RemoveContainer" containerID="a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.614553 4876 scope.go:117] "RemoveContainer" containerID="315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5" Mar 13 07:56:23 crc kubenswrapper[4876]: E0313 07:56:23.615230 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5\": container with ID starting with 315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5 not found: ID does not exist" containerID="315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.615348 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5"} err="failed to get container status \"315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5\": rpc error: code = NotFound desc = could not find container \"315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5\": container with ID starting with 315ac7c978324c325146a451e88445edb1d4359793637bcfb10b4d084f07c4e5 not found: ID does not exist" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.615465 4876 scope.go:117] "RemoveContainer" containerID="ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58" Mar 13 07:56:23 crc kubenswrapper[4876]: E0313 07:56:23.616176 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58\": container with ID starting with ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58 not found: ID does not exist" containerID="ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.616384 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58"} err="failed to get container status \"ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58\": rpc error: code = NotFound desc = could not find container \"ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58\": container with ID starting with ae18a74fcddd8c95ff7aad037ddd50b41af98e2e1aa8a5142f0d3e39df904c58 not found: ID does not exist" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.616486 4876 scope.go:117] "RemoveContainer" containerID="a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05" Mar 13 07:56:23 crc kubenswrapper[4876]: E0313 07:56:23.617013 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05\": container with ID starting with a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05 not found: ID does not exist" containerID="a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.617073 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05"} err="failed to get container status \"a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05\": rpc error: code = NotFound desc = could not find container \"a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05\": container with ID starting with a35a60a0dc01e90710fdb69aeeac6d1bdf9b8197cdbe101ae896a8f9c66edc05 not found: ID does not exist" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.796575 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-m7zdz" Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.807230 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nfcfw"] Mar 13 07:56:23 crc kubenswrapper[4876]: I0313 07:56:23.818774 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nfcfw"] Mar 13 07:56:24 crc kubenswrapper[4876]: I0313 07:56:24.607635 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:56:24 crc kubenswrapper[4876]: I0313 07:56:24.608315 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:56:24 crc kubenswrapper[4876]: I0313 07:56:24.608399 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:56:24 crc kubenswrapper[4876]: I0313 07:56:24.609416 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"82566a3994e302f89937ff7ed8edf05f5c82529ba0688b6946e78bb6b8475402"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 07:56:24 crc kubenswrapper[4876]: I0313 07:56:24.609533 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://82566a3994e302f89937ff7ed8edf05f5c82529ba0688b6946e78bb6b8475402" gracePeriod=600 Mar 13 07:56:24 crc kubenswrapper[4876]: E0313 07:56:24.774774 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a6f71e5_2091_4386_b559_bba70bc45972.slice/crio-82566a3994e302f89937ff7ed8edf05f5c82529ba0688b6946e78bb6b8475402.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a6f71e5_2091_4386_b559_bba70bc45972.slice/crio-conmon-82566a3994e302f89937ff7ed8edf05f5c82529ba0688b6946e78bb6b8475402.scope\": RecentStats: unable to find data in memory cache]" Mar 13 07:56:25 crc kubenswrapper[4876]: I0313 07:56:25.047734 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" path="/var/lib/kubelet/pods/6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb/volumes" Mar 13 07:56:25 crc kubenswrapper[4876]: I0313 07:56:25.048811 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" path="/var/lib/kubelet/pods/828564d0-2f02-42ba-a4b7-0ba9e06c987c/volumes" Mar 13 07:56:25 crc kubenswrapper[4876]: I0313 07:56:25.492898 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="82566a3994e302f89937ff7ed8edf05f5c82529ba0688b6946e78bb6b8475402" exitCode=0 Mar 13 07:56:25 crc kubenswrapper[4876]: I0313 07:56:25.492986 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"82566a3994e302f89937ff7ed8edf05f5c82529ba0688b6946e78bb6b8475402"} Mar 13 07:56:25 crc kubenswrapper[4876]: I0313 07:56:25.493069 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"8da8f1dc197e5903f65bce5f57eb064acd43a39afee0ba7b7e356d5bf1d3a13e"} Mar 13 07:56:25 crc kubenswrapper[4876]: I0313 07:56:25.493091 4876 scope.go:117] "RemoveContainer" containerID="8ebde89ffdd3ea68ea7599ed04541e15278a60938a622659662a7eee8835605e" Mar 13 07:56:33 crc kubenswrapper[4876]: I0313 07:56:33.337037 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-pfswh" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.703273 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h74p7"] Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.704860 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="extract-content" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.704883 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="extract-content" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.704926 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="extract-utilities" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.704935 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="extract-utilities" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.704952 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.704961 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.704974 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.704981 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.705010 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="extract-content" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.705018 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="extract-content" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.705027 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9431e26b-bb79-45a0-a0a2-9e11479cc64b" containerName="oc" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.705035 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9431e26b-bb79-45a0-a0a2-9e11479cc64b" containerName="oc" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.705045 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerName="extract-utilities" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.705052 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerName="extract-utilities" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.710330 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerName="extract-content" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.710394 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerName="extract-content" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.710441 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.710449 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: E0313 07:56:49.710462 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="extract-utilities" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.710470 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="extract-utilities" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.710917 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="74fa4506-343c-48a2-bb72-612cc2315daa" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.710949 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9431e26b-bb79-45a0-a0a2-9e11479cc64b" containerName="oc" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.710961 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fee2ff8-a4a7-4b69-96b5-b3a8ca5552cb" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.710991 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="828564d0-2f02-42ba-a4b7-0ba9e06c987c" containerName="registry-server" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.712299 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.717939 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h74p7"] Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.719139 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.719300 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.723566 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wr2zg" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.724085 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.790565 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3033f53f-f48e-4010-b13a-8cdd540a4a81-config\") pod \"dnsmasq-dns-675f4bcbfc-h74p7\" (UID: \"3033f53f-f48e-4010-b13a-8cdd540a4a81\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.790644 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw84b\" (UniqueName: \"kubernetes.io/projected/3033f53f-f48e-4010-b13a-8cdd540a4a81-kube-api-access-bw84b\") pod \"dnsmasq-dns-675f4bcbfc-h74p7\" (UID: \"3033f53f-f48e-4010-b13a-8cdd540a4a81\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.802050 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zn9kq"] Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.803305 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.805047 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.810190 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zn9kq"] Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.892495 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.892551 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw84b\" (UniqueName: \"kubernetes.io/projected/3033f53f-f48e-4010-b13a-8cdd540a4a81-kube-api-access-bw84b\") pod \"dnsmasq-dns-675f4bcbfc-h74p7\" (UID: \"3033f53f-f48e-4010-b13a-8cdd540a4a81\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.892617 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3033f53f-f48e-4010-b13a-8cdd540a4a81-config\") pod \"dnsmasq-dns-675f4bcbfc-h74p7\" (UID: \"3033f53f-f48e-4010-b13a-8cdd540a4a81\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.892640 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9nlx\" (UniqueName: \"kubernetes.io/projected/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-kube-api-access-l9nlx\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.892661 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-config\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.894074 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3033f53f-f48e-4010-b13a-8cdd540a4a81-config\") pod \"dnsmasq-dns-675f4bcbfc-h74p7\" (UID: \"3033f53f-f48e-4010-b13a-8cdd540a4a81\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.916781 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw84b\" (UniqueName: \"kubernetes.io/projected/3033f53f-f48e-4010-b13a-8cdd540a4a81-kube-api-access-bw84b\") pod \"dnsmasq-dns-675f4bcbfc-h74p7\" (UID: \"3033f53f-f48e-4010-b13a-8cdd540a4a81\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.993603 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9nlx\" (UniqueName: \"kubernetes.io/projected/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-kube-api-access-l9nlx\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.994455 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-config\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.994592 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.995548 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-config\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:49 crc kubenswrapper[4876]: I0313 07:56:49.995567 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:50 crc kubenswrapper[4876]: I0313 07:56:50.012756 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9nlx\" (UniqueName: \"kubernetes.io/projected/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-kube-api-access-l9nlx\") pod \"dnsmasq-dns-78dd6ddcc-zn9kq\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:50 crc kubenswrapper[4876]: I0313 07:56:50.037961 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:56:50 crc kubenswrapper[4876]: I0313 07:56:50.116409 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:56:50 crc kubenswrapper[4876]: I0313 07:56:50.419231 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zn9kq"] Mar 13 07:56:50 crc kubenswrapper[4876]: I0313 07:56:50.507605 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h74p7"] Mar 13 07:56:50 crc kubenswrapper[4876]: I0313 07:56:50.738835 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" event={"ID":"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e","Type":"ContainerStarted","Data":"1e05a85e7b6b1c1f8f75577f62eca7533ec8887d7550750ac2f9b557832617e5"} Mar 13 07:56:50 crc kubenswrapper[4876]: I0313 07:56:50.740340 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" event={"ID":"3033f53f-f48e-4010-b13a-8cdd540a4a81","Type":"ContainerStarted","Data":"b5ae78f5ef3029612a1483260e65631dc7a296a2cebe046479a32b661dbdc82f"} Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.616333 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h74p7"] Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.646997 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-fghn2"] Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.649391 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.662913 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-fghn2"] Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.761413 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxtn9\" (UniqueName: \"kubernetes.io/projected/a3ede6ca-f991-4143-885e-31fea503a54f-kube-api-access-lxtn9\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.761507 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-config\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.761531 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.874367 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxtn9\" (UniqueName: \"kubernetes.io/projected/a3ede6ca-f991-4143-885e-31fea503a54f-kube-api-access-lxtn9\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.874432 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-config\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.874450 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.875398 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.875574 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-config\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.930022 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxtn9\" (UniqueName: \"kubernetes.io/projected/a3ede6ca-f991-4143-885e-31fea503a54f-kube-api-access-lxtn9\") pod \"dnsmasq-dns-5ccc8479f9-fghn2\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.990838 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zn9kq"] Mar 13 07:56:52 crc kubenswrapper[4876]: I0313 07:56:52.996128 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.032290 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-w8nvw"] Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.039179 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.058151 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-w8nvw"] Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.079833 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-config\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.079930 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbp2h\" (UniqueName: \"kubernetes.io/projected/511f655c-0499-44c9-a654-0094ba185554-kube-api-access-cbp2h\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.079974 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.182341 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-config\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.182455 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbp2h\" (UniqueName: \"kubernetes.io/projected/511f655c-0499-44c9-a654-0094ba185554-kube-api-access-cbp2h\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.182499 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.183433 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.183633 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-config\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.235351 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbp2h\" (UniqueName: \"kubernetes.io/projected/511f655c-0499-44c9-a654-0094ba185554-kube-api-access-cbp2h\") pod \"dnsmasq-dns-57d769cc4f-w8nvw\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.368886 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.463930 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-fghn2"] Mar 13 07:56:53 crc kubenswrapper[4876]: W0313 07:56:53.477137 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3ede6ca_f991_4143_885e_31fea503a54f.slice/crio-1f715bdfb80664a058be63504a07a3cdc1fba229fe3e8e3113c9c0088f2c92d0 WatchSource:0}: Error finding container 1f715bdfb80664a058be63504a07a3cdc1fba229fe3e8e3113c9c0088f2c92d0: Status 404 returned error can't find the container with id 1f715bdfb80664a058be63504a07a3cdc1fba229fe3e8e3113c9c0088f2c92d0 Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.784702 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" event={"ID":"a3ede6ca-f991-4143-885e-31fea503a54f","Type":"ContainerStarted","Data":"1f715bdfb80664a058be63504a07a3cdc1fba229fe3e8e3113c9c0088f2c92d0"} Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.825593 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.830407 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.836158 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.836458 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.836626 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.836698 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.836626 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.836822 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.837015 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-csqv6" Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.855888 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 07:56:53 crc kubenswrapper[4876]: I0313 07:56:53.913529 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-w8nvw"] Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.001881 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.001933 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.001957 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.001980 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cca33d7d-7758-4f14-9d61-1785e84ea2b8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.002011 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cca33d7d-7758-4f14-9d61-1785e84ea2b8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.002039 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.002069 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.002086 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzcmk\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-kube-api-access-bzcmk\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.002545 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.002690 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.002757 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104200 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104308 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104340 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104389 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104413 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104445 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104470 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cca33d7d-7758-4f14-9d61-1785e84ea2b8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104505 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cca33d7d-7758-4f14-9d61-1785e84ea2b8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104535 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104562 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.104583 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzcmk\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-kube-api-access-bzcmk\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.105065 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.106154 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.106952 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.107091 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.107333 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.107915 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.112774 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.112874 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cca33d7d-7758-4f14-9d61-1785e84ea2b8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.113409 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cca33d7d-7758-4f14-9d61-1785e84ea2b8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.120697 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzcmk\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-kube-api-access-bzcmk\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.137455 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.138828 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.160572 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.186229 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.196288 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.205179 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.205269 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.205430 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2qq6k" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.205599 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.205769 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.205825 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.206020 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.257895 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.309979 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310080 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-server-conf\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310144 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310166 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310270 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310301 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310367 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310401 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j7p4\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-kube-api-access-4j7p4\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310470 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310500 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-pod-info\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.310537 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-config-data\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413228 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413306 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-server-conf\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413360 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413380 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413431 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413456 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413478 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413527 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j7p4\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-kube-api-access-4j7p4\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413559 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413601 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-pod-info\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.413634 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-config-data\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.414921 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-config-data\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.415659 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.416520 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-server-conf\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.417334 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.418838 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.418860 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.422514 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.442230 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j7p4\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-kube-api-access-4j7p4\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.447072 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.447582 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-pod-info\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.456615 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.480175 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.533123 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 07:56:54 crc kubenswrapper[4876]: I0313 07:56:54.827990 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" event={"ID":"511f655c-0499-44c9-a654-0094ba185554","Type":"ContainerStarted","Data":"a2199d5f27431fda9642b7cbd6afbbe1bae79e526ba8d23cde85f2b99e02ee7b"} Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.084821 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.090215 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.095773 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.096059 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-mp8ch" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.096676 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.096736 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.101254 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.117505 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.232425 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e32caff-0034-47f2-aa5b-917607dd44b6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.232489 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz44v\" (UniqueName: \"kubernetes.io/projected/7e32caff-0034-47f2-aa5b-917607dd44b6-kube-api-access-wz44v\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.232528 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.232585 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e32caff-0034-47f2-aa5b-917607dd44b6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.232615 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-config-data-default\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.232644 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.232689 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7e32caff-0034-47f2-aa5b-917607dd44b6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.232716 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-kolla-config\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.341457 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e32caff-0034-47f2-aa5b-917607dd44b6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.341522 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz44v\" (UniqueName: \"kubernetes.io/projected/7e32caff-0034-47f2-aa5b-917607dd44b6-kube-api-access-wz44v\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.341571 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.341633 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e32caff-0034-47f2-aa5b-917607dd44b6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.341686 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-config-data-default\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.341761 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.341805 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7e32caff-0034-47f2-aa5b-917607dd44b6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.341850 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-kolla-config\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.342652 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-kolla-config\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.343529 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-config-data-default\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.343837 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7e32caff-0034-47f2-aa5b-917607dd44b6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.344091 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.344991 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e32caff-0034-47f2-aa5b-917607dd44b6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.352520 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e32caff-0034-47f2-aa5b-917607dd44b6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.353026 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e32caff-0034-47f2-aa5b-917607dd44b6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.398596 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz44v\" (UniqueName: \"kubernetes.io/projected/7e32caff-0034-47f2-aa5b-917607dd44b6-kube-api-access-wz44v\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.413745 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"7e32caff-0034-47f2-aa5b-917607dd44b6\") " pod="openstack/openstack-galera-0" Mar 13 07:56:55 crc kubenswrapper[4876]: I0313 07:56:55.423901 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.511533 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.523582 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.546215 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hbcqh" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.548449 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.549110 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.550213 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.569378 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.723116 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.723184 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c36ecf-187d-4388-8d6f-626fe6470eb0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.723209 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e8c36ecf-187d-4388-8d6f-626fe6470eb0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.723271 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.723341 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.723363 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crrj\" (UniqueName: \"kubernetes.io/projected/e8c36ecf-187d-4388-8d6f-626fe6470eb0-kube-api-access-2crrj\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.723395 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c36ecf-187d-4388-8d6f-626fe6470eb0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.723417 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.824772 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.824913 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.824953 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crrj\" (UniqueName: \"kubernetes.io/projected/e8c36ecf-187d-4388-8d6f-626fe6470eb0-kube-api-access-2crrj\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.824977 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c36ecf-187d-4388-8d6f-626fe6470eb0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.825007 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.825043 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.825072 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c36ecf-187d-4388-8d6f-626fe6470eb0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.825102 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e8c36ecf-187d-4388-8d6f-626fe6470eb0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.825210 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.826529 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.826764 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e8c36ecf-187d-4388-8d6f-626fe6470eb0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.827054 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.829776 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e8c36ecf-187d-4388-8d6f-626fe6470eb0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.842078 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c36ecf-187d-4388-8d6f-626fe6470eb0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.847963 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c36ecf-187d-4388-8d6f-626fe6470eb0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.848042 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crrj\" (UniqueName: \"kubernetes.io/projected/e8c36ecf-187d-4388-8d6f-626fe6470eb0-kube-api-access-2crrj\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.855820 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.857229 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.865633 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e8c36ecf-187d-4388-8d6f-626fe6470eb0\") " pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.867582 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.872010 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-7svkt" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.874205 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.883261 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 07:56:56 crc kubenswrapper[4876]: I0313 07:56:56.892267 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.028648 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e221cd11-36ee-4c17-a26d-e95b75ed5682-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.028764 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e221cd11-36ee-4c17-a26d-e95b75ed5682-kolla-config\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.028810 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e221cd11-36ee-4c17-a26d-e95b75ed5682-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.028857 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e221cd11-36ee-4c17-a26d-e95b75ed5682-config-data\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.028886 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v9zd\" (UniqueName: \"kubernetes.io/projected/e221cd11-36ee-4c17-a26d-e95b75ed5682-kube-api-access-2v9zd\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.131173 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e221cd11-36ee-4c17-a26d-e95b75ed5682-kolla-config\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.131261 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e221cd11-36ee-4c17-a26d-e95b75ed5682-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.131324 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e221cd11-36ee-4c17-a26d-e95b75ed5682-config-data\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.131353 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v9zd\" (UniqueName: \"kubernetes.io/projected/e221cd11-36ee-4c17-a26d-e95b75ed5682-kube-api-access-2v9zd\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.131385 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e221cd11-36ee-4c17-a26d-e95b75ed5682-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.132125 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e221cd11-36ee-4c17-a26d-e95b75ed5682-kolla-config\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.133024 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e221cd11-36ee-4c17-a26d-e95b75ed5682-config-data\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.137358 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e221cd11-36ee-4c17-a26d-e95b75ed5682-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.141008 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e221cd11-36ee-4c17-a26d-e95b75ed5682-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.158777 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v9zd\" (UniqueName: \"kubernetes.io/projected/e221cd11-36ee-4c17-a26d-e95b75ed5682-kube-api-access-2v9zd\") pod \"memcached-0\" (UID: \"e221cd11-36ee-4c17-a26d-e95b75ed5682\") " pod="openstack/memcached-0" Mar 13 07:56:57 crc kubenswrapper[4876]: I0313 07:56:57.221542 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 07:56:59 crc kubenswrapper[4876]: I0313 07:56:59.247287 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 07:56:59 crc kubenswrapper[4876]: I0313 07:56:59.249535 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 07:56:59 crc kubenswrapper[4876]: I0313 07:56:59.251984 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-bv97s" Mar 13 07:56:59 crc kubenswrapper[4876]: I0313 07:56:59.268749 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 07:56:59 crc kubenswrapper[4876]: I0313 07:56:59.365105 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69g2n\" (UniqueName: \"kubernetes.io/projected/7a6b118f-c518-4e1e-9f0a-79bb844a304f-kube-api-access-69g2n\") pod \"kube-state-metrics-0\" (UID: \"7a6b118f-c518-4e1e-9f0a-79bb844a304f\") " pod="openstack/kube-state-metrics-0" Mar 13 07:56:59 crc kubenswrapper[4876]: I0313 07:56:59.466760 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69g2n\" (UniqueName: \"kubernetes.io/projected/7a6b118f-c518-4e1e-9f0a-79bb844a304f-kube-api-access-69g2n\") pod \"kube-state-metrics-0\" (UID: \"7a6b118f-c518-4e1e-9f0a-79bb844a304f\") " pod="openstack/kube-state-metrics-0" Mar 13 07:56:59 crc kubenswrapper[4876]: I0313 07:56:59.499912 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69g2n\" (UniqueName: \"kubernetes.io/projected/7a6b118f-c518-4e1e-9f0a-79bb844a304f-kube-api-access-69g2n\") pod \"kube-state-metrics-0\" (UID: \"7a6b118f-c518-4e1e-9f0a-79bb844a304f\") " pod="openstack/kube-state-metrics-0" Mar 13 07:56:59 crc kubenswrapper[4876]: I0313 07:56:59.570620 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 07:57:01 crc kubenswrapper[4876]: I0313 07:57:01.854278 4876 scope.go:117] "RemoveContainer" containerID="89186f77092dbe556c0d7aa5a6c5c92db61dc352f103ab14880e73967e39e98c" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.649710 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xrn6k"] Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.651178 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.655560 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-xprp4" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.655798 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.661546 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.669723 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xrn6k"] Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.686554 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-tb4jh"] Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.688906 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.700852 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-tb4jh"] Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.747598 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b55481b-c5a5-4602-8f66-c3b002134edb-ovn-controller-tls-certs\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.747985 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b55481b-c5a5-4602-8f66-c3b002134edb-combined-ca-bundle\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.748100 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-run\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.748182 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b55481b-c5a5-4602-8f66-c3b002134edb-scripts\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.748534 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-log-ovn\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.748635 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q75lw\" (UniqueName: \"kubernetes.io/projected/3b55481b-c5a5-4602-8f66-c3b002134edb-kube-api-access-q75lw\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.748727 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-run-ovn\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.849985 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b55481b-c5a5-4602-8f66-c3b002134edb-ovn-controller-tls-certs\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850087 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-etc-ovs\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850173 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b55481b-c5a5-4602-8f66-c3b002134edb-combined-ca-bundle\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850218 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsg6f\" (UniqueName: \"kubernetes.io/projected/23f73ce7-85a0-410c-a3f3-f1df187b509e-kube-api-access-rsg6f\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850323 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-log\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850367 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-run\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850407 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b55481b-c5a5-4602-8f66-c3b002134edb-scripts\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850448 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-log-ovn\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850479 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q75lw\" (UniqueName: \"kubernetes.io/projected/3b55481b-c5a5-4602-8f66-c3b002134edb-kube-api-access-q75lw\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850585 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-run-ovn\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850630 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23f73ce7-85a0-410c-a3f3-f1df187b509e-scripts\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850681 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-run\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.850716 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-lib\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.851719 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-run\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.852273 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-run-ovn\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.852407 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3b55481b-c5a5-4602-8f66-c3b002134edb-var-log-ovn\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.855589 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b55481b-c5a5-4602-8f66-c3b002134edb-scripts\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.857102 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b55481b-c5a5-4602-8f66-c3b002134edb-ovn-controller-tls-certs\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.858646 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b55481b-c5a5-4602-8f66-c3b002134edb-combined-ca-bundle\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.884927 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q75lw\" (UniqueName: \"kubernetes.io/projected/3b55481b-c5a5-4602-8f66-c3b002134edb-kube-api-access-q75lw\") pod \"ovn-controller-xrn6k\" (UID: \"3b55481b-c5a5-4602-8f66-c3b002134edb\") " pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.913936 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.915450 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.917946 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.917946 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-ds2dl" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.918203 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.918290 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.918472 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.930433 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.952694 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsg6f\" (UniqueName: \"kubernetes.io/projected/23f73ce7-85a0-410c-a3f3-f1df187b509e-kube-api-access-rsg6f\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.952758 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-log\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.953176 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23f73ce7-85a0-410c-a3f3-f1df187b509e-scripts\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.953409 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-log\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.954015 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-run\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.954075 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-lib\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.954182 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-run\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.954281 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-etc-ovs\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.954492 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-var-lib\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.954658 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/23f73ce7-85a0-410c-a3f3-f1df187b509e-etc-ovs\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.963026 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23f73ce7-85a0-410c-a3f3-f1df187b509e-scripts\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.970814 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsg6f\" (UniqueName: \"kubernetes.io/projected/23f73ce7-85a0-410c-a3f3-f1df187b509e-kube-api-access-rsg6f\") pod \"ovn-controller-ovs-tb4jh\" (UID: \"23f73ce7-85a0-410c-a3f3-f1df187b509e\") " pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:02 crc kubenswrapper[4876]: I0313 07:57:02.971358 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.012107 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.056174 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88e4b876-8c78-4d88-945b-bd10555503d2-config\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.056323 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.056361 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.056390 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.056420 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/88e4b876-8c78-4d88-945b-bd10555503d2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.056439 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.056468 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rqph\" (UniqueName: \"kubernetes.io/projected/88e4b876-8c78-4d88-945b-bd10555503d2-kube-api-access-2rqph\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.056519 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88e4b876-8c78-4d88-945b-bd10555503d2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.159223 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88e4b876-8c78-4d88-945b-bd10555503d2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.159364 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88e4b876-8c78-4d88-945b-bd10555503d2-config\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.159403 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.159437 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.159466 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.159494 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/88e4b876-8c78-4d88-945b-bd10555503d2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.159512 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.159542 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rqph\" (UniqueName: \"kubernetes.io/projected/88e4b876-8c78-4d88-945b-bd10555503d2-kube-api-access-2rqph\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.161219 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88e4b876-8c78-4d88-945b-bd10555503d2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.169008 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.169678 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88e4b876-8c78-4d88-945b-bd10555503d2-config\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.169932 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/88e4b876-8c78-4d88-945b-bd10555503d2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.169928 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.176620 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.180662 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e4b876-8c78-4d88-945b-bd10555503d2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.201282 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rqph\" (UniqueName: \"kubernetes.io/projected/88e4b876-8c78-4d88-945b-bd10555503d2-kube-api-access-2rqph\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.227918 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"88e4b876-8c78-4d88-945b-bd10555503d2\") " pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:03 crc kubenswrapper[4876]: I0313 07:57:03.251935 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.137073 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.138970 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.146979 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.147538 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.147840 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-smnn6" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.152371 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.166807 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.242566 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.242656 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.242692 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.242722 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.242868 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.242957 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.243116 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-config\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.243153 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gngjf\" (UniqueName: \"kubernetes.io/projected/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-kube-api-access-gngjf\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.345501 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-config\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.345562 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gngjf\" (UniqueName: \"kubernetes.io/projected/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-kube-api-access-gngjf\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.345636 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.345666 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.345851 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.345880 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.345908 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.345938 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.346536 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-config\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.346866 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.348158 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.349023 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.357879 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.357900 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.358102 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.372162 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.373380 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gngjf\" (UniqueName: \"kubernetes.io/projected/37b39bdf-9f1c-4f03-88e5-6abd3949fc7c-kube-api-access-gngjf\") pod \"ovsdbserver-sb-0\" (UID: \"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c\") " pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:06 crc kubenswrapper[4876]: I0313 07:57:06.485504 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.205503 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.206278 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bw84b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-h74p7_openstack(3033f53f-f48e-4010-b13a-8cdd540a4a81): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.207481 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" podUID="3033f53f-f48e-4010-b13a-8cdd540a4a81" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.272223 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.272460 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lxtn9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-fghn2_openstack(a3ede6ca-f991-4143-885e-31fea503a54f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.275331 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" podUID="a3ede6ca-f991-4143-885e-31fea503a54f" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.303932 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.304136 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l9nlx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-zn9kq_openstack(c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:57:14 crc kubenswrapper[4876]: E0313 07:57:14.305376 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" podUID="c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e" Mar 13 07:57:14 crc kubenswrapper[4876]: I0313 07:57:14.755016 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 07:57:14 crc kubenswrapper[4876]: W0313 07:57:14.762273 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a6b118f_c518_4e1e_9f0a_79bb844a304f.slice/crio-ed426a1f40e97f2e7e362ae43bc3ee56c7291153e7af4a7ed2a58dbb1db7e12f WatchSource:0}: Error finding container ed426a1f40e97f2e7e362ae43bc3ee56c7291153e7af4a7ed2a58dbb1db7e12f: Status 404 returned error can't find the container with id ed426a1f40e97f2e7e362ae43bc3ee56c7291153e7af4a7ed2a58dbb1db7e12f Mar 13 07:57:14 crc kubenswrapper[4876]: I0313 07:57:14.763701 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 07:57:14 crc kubenswrapper[4876]: W0313 07:57:14.776851 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e32caff_0034_47f2_aa5b_917607dd44b6.slice/crio-30e2151edb1e5fe906183186b169072a0e16f7a94bcb5e2e10347a4412a14c3e WatchSource:0}: Error finding container 30e2151edb1e5fe906183186b169072a0e16f7a94bcb5e2e10347a4412a14c3e: Status 404 returned error can't find the container with id 30e2151edb1e5fe906183186b169072a0e16f7a94bcb5e2e10347a4412a14c3e Mar 13 07:57:14 crc kubenswrapper[4876]: I0313 07:57:14.941295 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 07:57:14 crc kubenswrapper[4876]: W0313 07:57:14.945057 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8c36ecf_187d_4388_8d6f_626fe6470eb0.slice/crio-5d6e7fca3c98ef41a0648bab95ae58d6baee7f6ddc26cfd7cbdf9745b7fa5a8f WatchSource:0}: Error finding container 5d6e7fca3c98ef41a0648bab95ae58d6baee7f6ddc26cfd7cbdf9745b7fa5a8f: Status 404 returned error can't find the container with id 5d6e7fca3c98ef41a0648bab95ae58d6baee7f6ddc26cfd7cbdf9745b7fa5a8f Mar 13 07:57:14 crc kubenswrapper[4876]: I0313 07:57:14.960600 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 07:57:14 crc kubenswrapper[4876]: I0313 07:57:14.975644 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.081820 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e8c36ecf-187d-4388-8d6f-626fe6470eb0","Type":"ContainerStarted","Data":"5d6e7fca3c98ef41a0648bab95ae58d6baee7f6ddc26cfd7cbdf9745b7fa5a8f"} Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.083493 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7e32caff-0034-47f2-aa5b-917607dd44b6","Type":"ContainerStarted","Data":"30e2151edb1e5fe906183186b169072a0e16f7a94bcb5e2e10347a4412a14c3e"} Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.084718 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e221cd11-36ee-4c17-a26d-e95b75ed5682","Type":"ContainerStarted","Data":"7a694a844d24759df36f263844cbee24e944d20a16d6e0a0319c1a01fe7793e6"} Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.085759 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7a6b118f-c518-4e1e-9f0a-79bb844a304f","Type":"ContainerStarted","Data":"ed426a1f40e97f2e7e362ae43bc3ee56c7291153e7af4a7ed2a58dbb1db7e12f"} Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.089440 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678","Type":"ContainerStarted","Data":"3773a980a2d5a23e03d1ec13efadf8a908102dd731491440f7d484762ca9d9d0"} Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.091729 4876 generic.go:334] "Generic (PLEG): container finished" podID="511f655c-0499-44c9-a654-0094ba185554" containerID="d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d" exitCode=0 Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.093299 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" event={"ID":"511f655c-0499-44c9-a654-0094ba185554","Type":"ContainerDied","Data":"d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d"} Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.215255 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.225604 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xrn6k"] Mar 13 07:57:15 crc kubenswrapper[4876]: W0313 07:57:15.258636 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcca33d7d_7758_4f14_9d61_1785e84ea2b8.slice/crio-d0d93982bd721aa0e0c61c7c3bcd60b435563cc240c52a9857df0cf9d4371b60 WatchSource:0}: Error finding container d0d93982bd721aa0e0c61c7c3bcd60b435563cc240c52a9857df0cf9d4371b60: Status 404 returned error can't find the container with id d0d93982bd721aa0e0c61c7c3bcd60b435563cc240c52a9857df0cf9d4371b60 Mar 13 07:57:15 crc kubenswrapper[4876]: W0313 07:57:15.268649 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b55481b_c5a5_4602_8f66_c3b002134edb.slice/crio-6d0791656104c31051e83c98ced29587a6f082ebdad64f5c55424acec3b8fcb4 WatchSource:0}: Error finding container 6d0791656104c31051e83c98ced29587a6f082ebdad64f5c55424acec3b8fcb4: Status 404 returned error can't find the container with id 6d0791656104c31051e83c98ced29587a6f082ebdad64f5c55424acec3b8fcb4 Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.409475 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.473786 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 07:57:15 crc kubenswrapper[4876]: W0313 07:57:15.582404 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88e4b876_8c78_4d88_945b_bd10555503d2.slice/crio-abbc409a8ae230c34163243b6a7615ea59c01acae23c57b5e4ea38c83951f0e1 WatchSource:0}: Error finding container abbc409a8ae230c34163243b6a7615ea59c01acae23c57b5e4ea38c83951f0e1: Status 404 returned error can't find the container with id abbc409a8ae230c34163243b6a7615ea59c01acae23c57b5e4ea38c83951f0e1 Mar 13 07:57:15 crc kubenswrapper[4876]: E0313 07:57:15.638993 4876 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Mar 13 07:57:15 crc kubenswrapper[4876]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/a3ede6ca-f991-4143-885e-31fea503a54f/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 13 07:57:15 crc kubenswrapper[4876]: > podSandboxID="1f715bdfb80664a058be63504a07a3cdc1fba229fe3e8e3113c9c0088f2c92d0" Mar 13 07:57:15 crc kubenswrapper[4876]: E0313 07:57:15.639657 4876 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 07:57:15 crc kubenswrapper[4876]: init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lxtn9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-fghn2_openstack(a3ede6ca-f991-4143-885e-31fea503a54f): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/a3ede6ca-f991-4143-885e-31fea503a54f/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 13 07:57:15 crc kubenswrapper[4876]: > logger="UnhandledError" Mar 13 07:57:15 crc kubenswrapper[4876]: E0313 07:57:15.641204 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/a3ede6ca-f991-4143-885e-31fea503a54f/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" podUID="a3ede6ca-f991-4143-885e-31fea503a54f" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.659782 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.744151 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-config\") pod \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.744359 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-dns-svc\") pod \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.744389 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9nlx\" (UniqueName: \"kubernetes.io/projected/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-kube-api-access-l9nlx\") pod \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\" (UID: \"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e\") " Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.746216 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-config" (OuterVolumeSpecName: "config") pod "c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e" (UID: "c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.746330 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e" (UID: "c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.755459 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-kube-api-access-l9nlx" (OuterVolumeSpecName: "kube-api-access-l9nlx") pod "c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e" (UID: "c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e"). InnerVolumeSpecName "kube-api-access-l9nlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.781111 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.846174 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw84b\" (UniqueName: \"kubernetes.io/projected/3033f53f-f48e-4010-b13a-8cdd540a4a81-kube-api-access-bw84b\") pod \"3033f53f-f48e-4010-b13a-8cdd540a4a81\" (UID: \"3033f53f-f48e-4010-b13a-8cdd540a4a81\") " Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.846507 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3033f53f-f48e-4010-b13a-8cdd540a4a81-config\") pod \"3033f53f-f48e-4010-b13a-8cdd540a4a81\" (UID: \"3033f53f-f48e-4010-b13a-8cdd540a4a81\") " Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.846923 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.846989 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.847073 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9nlx\" (UniqueName: \"kubernetes.io/projected/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e-kube-api-access-l9nlx\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.847632 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3033f53f-f48e-4010-b13a-8cdd540a4a81-config" (OuterVolumeSpecName: "config") pod "3033f53f-f48e-4010-b13a-8cdd540a4a81" (UID: "3033f53f-f48e-4010-b13a-8cdd540a4a81"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.850952 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3033f53f-f48e-4010-b13a-8cdd540a4a81-kube-api-access-bw84b" (OuterVolumeSpecName: "kube-api-access-bw84b") pod "3033f53f-f48e-4010-b13a-8cdd540a4a81" (UID: "3033f53f-f48e-4010-b13a-8cdd540a4a81"). InnerVolumeSpecName "kube-api-access-bw84b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.949181 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw84b\" (UniqueName: \"kubernetes.io/projected/3033f53f-f48e-4010-b13a-8cdd540a4a81-kube-api-access-bw84b\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:15 crc kubenswrapper[4876]: I0313 07:57:15.949223 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3033f53f-f48e-4010-b13a-8cdd540a4a81-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.107140 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.107120 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zn9kq" event={"ID":"c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e","Type":"ContainerDied","Data":"1e05a85e7b6b1c1f8f75577f62eca7533ec8887d7550750ac2f9b557832617e5"} Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.110125 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xrn6k" event={"ID":"3b55481b-c5a5-4602-8f66-c3b002134edb","Type":"ContainerStarted","Data":"6d0791656104c31051e83c98ced29587a6f082ebdad64f5c55424acec3b8fcb4"} Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.113679 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" event={"ID":"511f655c-0499-44c9-a654-0094ba185554","Type":"ContainerStarted","Data":"b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61"} Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.113833 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.118125 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"88e4b876-8c78-4d88-945b-bd10555503d2","Type":"ContainerStarted","Data":"abbc409a8ae230c34163243b6a7615ea59c01acae23c57b5e4ea38c83951f0e1"} Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.120227 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.120226 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-h74p7" event={"ID":"3033f53f-f48e-4010-b13a-8cdd540a4a81","Type":"ContainerDied","Data":"b5ae78f5ef3029612a1483260e65631dc7a296a2cebe046479a32b661dbdc82f"} Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.122024 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cca33d7d-7758-4f14-9d61-1785e84ea2b8","Type":"ContainerStarted","Data":"d0d93982bd721aa0e0c61c7c3bcd60b435563cc240c52a9857df0cf9d4371b60"} Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.123730 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c","Type":"ContainerStarted","Data":"f74997e899bca84c59fa90d0afd7f684e859ea661479f51112b8290a8410430e"} Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.174570 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" podStartSLOduration=2.643162861 podStartE2EDuration="23.174541479s" podCreationTimestamp="2026-03-13 07:56:53 +0000 UTC" firstStartedPulling="2026-03-13 07:56:53.93508485 +0000 UTC m=+1073.605863822" lastFinishedPulling="2026-03-13 07:57:14.466463458 +0000 UTC m=+1094.137242440" observedRunningTime="2026-03-13 07:57:16.137831524 +0000 UTC m=+1095.808610496" watchObservedRunningTime="2026-03-13 07:57:16.174541479 +0000 UTC m=+1095.845320461" Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.183451 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zn9kq"] Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.197038 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zn9kq"] Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.265122 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h74p7"] Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.284033 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h74p7"] Mar 13 07:57:16 crc kubenswrapper[4876]: I0313 07:57:16.556200 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-tb4jh"] Mar 13 07:57:16 crc kubenswrapper[4876]: W0313 07:57:16.713418 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23f73ce7_85a0_410c_a3f3_f1df187b509e.slice/crio-97125f223415406b8fa85be5eb337c118213a5a9f0974f844e04587a90590afc WatchSource:0}: Error finding container 97125f223415406b8fa85be5eb337c118213a5a9f0974f844e04587a90590afc: Status 404 returned error can't find the container with id 97125f223415406b8fa85be5eb337c118213a5a9f0974f844e04587a90590afc Mar 13 07:57:17 crc kubenswrapper[4876]: I0313 07:57:17.048888 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3033f53f-f48e-4010-b13a-8cdd540a4a81" path="/var/lib/kubelet/pods/3033f53f-f48e-4010-b13a-8cdd540a4a81/volumes" Mar 13 07:57:17 crc kubenswrapper[4876]: I0313 07:57:17.049372 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e" path="/var/lib/kubelet/pods/c4d41a0e-b6ff-4e2b-8491-6df7e7b1994e/volumes" Mar 13 07:57:17 crc kubenswrapper[4876]: I0313 07:57:17.133517 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tb4jh" event={"ID":"23f73ce7-85a0-410c-a3f3-f1df187b509e","Type":"ContainerStarted","Data":"97125f223415406b8fa85be5eb337c118213a5a9f0974f844e04587a90590afc"} Mar 13 07:57:23 crc kubenswrapper[4876]: I0313 07:57:23.378420 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:57:23 crc kubenswrapper[4876]: I0313 07:57:23.450798 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-fghn2"] Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.228728 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" event={"ID":"a3ede6ca-f991-4143-885e-31fea503a54f","Type":"ContainerDied","Data":"1f715bdfb80664a058be63504a07a3cdc1fba229fe3e8e3113c9c0088f2c92d0"} Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.229587 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f715bdfb80664a058be63504a07a3cdc1fba229fe3e8e3113c9c0088f2c92d0" Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.233185 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.262723 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-config\") pod \"a3ede6ca-f991-4143-885e-31fea503a54f\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.262844 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-dns-svc\") pod \"a3ede6ca-f991-4143-885e-31fea503a54f\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.262968 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxtn9\" (UniqueName: \"kubernetes.io/projected/a3ede6ca-f991-4143-885e-31fea503a54f-kube-api-access-lxtn9\") pod \"a3ede6ca-f991-4143-885e-31fea503a54f\" (UID: \"a3ede6ca-f991-4143-885e-31fea503a54f\") " Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.269275 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3ede6ca-f991-4143-885e-31fea503a54f-kube-api-access-lxtn9" (OuterVolumeSpecName: "kube-api-access-lxtn9") pod "a3ede6ca-f991-4143-885e-31fea503a54f" (UID: "a3ede6ca-f991-4143-885e-31fea503a54f"). InnerVolumeSpecName "kube-api-access-lxtn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.303501 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a3ede6ca-f991-4143-885e-31fea503a54f" (UID: "a3ede6ca-f991-4143-885e-31fea503a54f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.312051 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-config" (OuterVolumeSpecName: "config") pod "a3ede6ca-f991-4143-885e-31fea503a54f" (UID: "a3ede6ca-f991-4143-885e-31fea503a54f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.365834 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxtn9\" (UniqueName: \"kubernetes.io/projected/a3ede6ca-f991-4143-885e-31fea503a54f-kube-api-access-lxtn9\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.365886 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:25 crc kubenswrapper[4876]: I0313 07:57:25.365897 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3ede6ca-f991-4143-885e-31fea503a54f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.240774 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"88e4b876-8c78-4d88-945b-bd10555503d2","Type":"ContainerStarted","Data":"d14880ea7202d0d045bf3f814f5b362b4b3d2b49829b632ab0101e670d6320f9"} Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.243533 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e8c36ecf-187d-4388-8d6f-626fe6470eb0","Type":"ContainerStarted","Data":"7f96021ae8960574047284450c84386f049a619d4e6b3f3a6b7b9c392c2d8af6"} Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.245607 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7a6b118f-c518-4e1e-9f0a-79bb844a304f","Type":"ContainerStarted","Data":"9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345"} Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.245843 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.248373 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-fghn2" Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.248384 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e221cd11-36ee-4c17-a26d-e95b75ed5682","Type":"ContainerStarted","Data":"2ec5ec46323ffc7c6517efd05cf8da8156b3ee66e9836beed1cef3262be7a45d"} Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.248600 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.323786 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-fghn2"] Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.374121 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-fghn2"] Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.379871 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.293203926 podStartE2EDuration="30.379836173s" podCreationTimestamp="2026-03-13 07:56:56 +0000 UTC" firstStartedPulling="2026-03-13 07:57:14.976264245 +0000 UTC m=+1094.647043227" lastFinishedPulling="2026-03-13 07:57:25.062896462 +0000 UTC m=+1104.733675474" observedRunningTime="2026-03-13 07:57:26.352201648 +0000 UTC m=+1106.022980630" watchObservedRunningTime="2026-03-13 07:57:26.379836173 +0000 UTC m=+1106.050615155" Mar 13 07:57:26 crc kubenswrapper[4876]: I0313 07:57:26.386427 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.988381667 podStartE2EDuration="27.38641556s" podCreationTimestamp="2026-03-13 07:56:59 +0000 UTC" firstStartedPulling="2026-03-13 07:57:14.767410863 +0000 UTC m=+1094.438189845" lastFinishedPulling="2026-03-13 07:57:25.165444756 +0000 UTC m=+1104.836223738" observedRunningTime="2026-03-13 07:57:26.370498728 +0000 UTC m=+1106.041277720" watchObservedRunningTime="2026-03-13 07:57:26.38641556 +0000 UTC m=+1106.057194542" Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.054618 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3ede6ca-f991-4143-885e-31fea503a54f" path="/var/lib/kubelet/pods/a3ede6ca-f991-4143-885e-31fea503a54f/volumes" Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.258619 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xrn6k" event={"ID":"3b55481b-c5a5-4602-8f66-c3b002134edb","Type":"ContainerStarted","Data":"cd7c52bc3723f0199318fcb48395abb9161ab323a8aa125fe538a562f2f98564"} Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.258749 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-xrn6k" Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.263039 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7e32caff-0034-47f2-aa5b-917607dd44b6","Type":"ContainerStarted","Data":"369ed44bbeaac114256bbca2940c9ef1838a2cce53aba0a56f75ef42542dc5d3"} Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.267774 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cca33d7d-7758-4f14-9d61-1785e84ea2b8","Type":"ContainerStarted","Data":"8bbb2cd4e0422ba5625aa003e75bef5ff7479286bc579a6d22675536aa504525"} Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.276677 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tb4jh" event={"ID":"23f73ce7-85a0-410c-a3f3-f1df187b509e","Type":"ContainerStarted","Data":"076c77ee92ffb39e8ad94166c8bb9bce1ee590640b06d36cf67dde81174554bd"} Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.282161 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xrn6k" podStartSLOduration=15.398303889 podStartE2EDuration="25.282133652s" podCreationTimestamp="2026-03-13 07:57:02 +0000 UTC" firstStartedPulling="2026-03-13 07:57:15.285561385 +0000 UTC m=+1094.956340367" lastFinishedPulling="2026-03-13 07:57:25.169391148 +0000 UTC m=+1104.840170130" observedRunningTime="2026-03-13 07:57:27.277985764 +0000 UTC m=+1106.948764756" watchObservedRunningTime="2026-03-13 07:57:27.282133652 +0000 UTC m=+1106.952912634" Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.283864 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c","Type":"ContainerStarted","Data":"7ab4931c117d079111d99f15a54b8b93e9aac772ca16bfcbc958cca11c591513"} Mar 13 07:57:27 crc kubenswrapper[4876]: I0313 07:57:27.287594 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678","Type":"ContainerStarted","Data":"7bef2f513a2c87a154de6467ad78491d761e7c372e49596ffdb8f85720c39dea"} Mar 13 07:57:28 crc kubenswrapper[4876]: I0313 07:57:28.298758 4876 generic.go:334] "Generic (PLEG): container finished" podID="23f73ce7-85a0-410c-a3f3-f1df187b509e" containerID="076c77ee92ffb39e8ad94166c8bb9bce1ee590640b06d36cf67dde81174554bd" exitCode=0 Mar 13 07:57:28 crc kubenswrapper[4876]: I0313 07:57:28.299638 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tb4jh" event={"ID":"23f73ce7-85a0-410c-a3f3-f1df187b509e","Type":"ContainerDied","Data":"076c77ee92ffb39e8ad94166c8bb9bce1ee590640b06d36cf67dde81174554bd"} Mar 13 07:57:30 crc kubenswrapper[4876]: I0313 07:57:30.324990 4876 generic.go:334] "Generic (PLEG): container finished" podID="e8c36ecf-187d-4388-8d6f-626fe6470eb0" containerID="7f96021ae8960574047284450c84386f049a619d4e6b3f3a6b7b9c392c2d8af6" exitCode=0 Mar 13 07:57:30 crc kubenswrapper[4876]: I0313 07:57:30.325106 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e8c36ecf-187d-4388-8d6f-626fe6470eb0","Type":"ContainerDied","Data":"7f96021ae8960574047284450c84386f049a619d4e6b3f3a6b7b9c392c2d8af6"} Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.336114 4876 generic.go:334] "Generic (PLEG): container finished" podID="7e32caff-0034-47f2-aa5b-917607dd44b6" containerID="369ed44bbeaac114256bbca2940c9ef1838a2cce53aba0a56f75ef42542dc5d3" exitCode=0 Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.336201 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7e32caff-0034-47f2-aa5b-917607dd44b6","Type":"ContainerDied","Data":"369ed44bbeaac114256bbca2940c9ef1838a2cce53aba0a56f75ef42542dc5d3"} Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.339695 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tb4jh" event={"ID":"23f73ce7-85a0-410c-a3f3-f1df187b509e","Type":"ContainerStarted","Data":"2a0dbf4ba2efbe4034f7ccd2e1736704e26abe5506d9fc9ccb50fa71e806a297"} Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.339739 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tb4jh" event={"ID":"23f73ce7-85a0-410c-a3f3-f1df187b509e","Type":"ContainerStarted","Data":"a266085225ad5644b7c8e07d7c25a1c448e106bc8e01e6bbe95ecf5ee0d6468a"} Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.340309 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.340339 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.341794 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37b39bdf-9f1c-4f03-88e5-6abd3949fc7c","Type":"ContainerStarted","Data":"8e7ab378c8bb141faf25c0a3086ee4697be310dda7e7d917191c65d160be4b8c"} Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.353617 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"88e4b876-8c78-4d88-945b-bd10555503d2","Type":"ContainerStarted","Data":"7b0abb8178fb486ace4b747d38974b6c114497ed90593af71e774ba405eaca9f"} Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.381037 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e8c36ecf-187d-4388-8d6f-626fe6470eb0","Type":"ContainerStarted","Data":"04083e30a8bb91cb85110e0826b2eeb8c85604e4a75722cfe51ad2253badd4cc"} Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.416016 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=15.921282258 podStartE2EDuration="30.415989747s" podCreationTimestamp="2026-03-13 07:57:01 +0000 UTC" firstStartedPulling="2026-03-13 07:57:15.585386878 +0000 UTC m=+1095.256165860" lastFinishedPulling="2026-03-13 07:57:30.080094337 +0000 UTC m=+1109.750873349" observedRunningTime="2026-03-13 07:57:31.408768732 +0000 UTC m=+1111.079547704" watchObservedRunningTime="2026-03-13 07:57:31.415989747 +0000 UTC m=+1111.086768729" Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.446126 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.96292006 podStartE2EDuration="26.446106273s" podCreationTimestamp="2026-03-13 07:57:05 +0000 UTC" firstStartedPulling="2026-03-13 07:57:15.59010371 +0000 UTC m=+1095.260882692" lastFinishedPulling="2026-03-13 07:57:30.073289913 +0000 UTC m=+1109.744068905" observedRunningTime="2026-03-13 07:57:31.439907846 +0000 UTC m=+1111.110686828" watchObservedRunningTime="2026-03-13 07:57:31.446106273 +0000 UTC m=+1111.116885255" Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.485766 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.487979 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-tb4jh" podStartSLOduration=21.034751956 podStartE2EDuration="29.487968012s" podCreationTimestamp="2026-03-13 07:57:02 +0000 UTC" firstStartedPulling="2026-03-13 07:57:16.715564704 +0000 UTC m=+1096.386343686" lastFinishedPulling="2026-03-13 07:57:25.16878076 +0000 UTC m=+1104.839559742" observedRunningTime="2026-03-13 07:57:31.479317616 +0000 UTC m=+1111.150096598" watchObservedRunningTime="2026-03-13 07:57:31.487968012 +0000 UTC m=+1111.158746994" Mar 13 07:57:31 crc kubenswrapper[4876]: I0313 07:57:31.510036 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.295633107 podStartE2EDuration="36.510004948s" podCreationTimestamp="2026-03-13 07:56:55 +0000 UTC" firstStartedPulling="2026-03-13 07:57:14.951184498 +0000 UTC m=+1094.621963480" lastFinishedPulling="2026-03-13 07:57:25.165556339 +0000 UTC m=+1104.836335321" observedRunningTime="2026-03-13 07:57:31.507910659 +0000 UTC m=+1111.178689651" watchObservedRunningTime="2026-03-13 07:57:31.510004948 +0000 UTC m=+1111.180783930" Mar 13 07:57:32 crc kubenswrapper[4876]: I0313 07:57:32.223552 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 13 07:57:32 crc kubenswrapper[4876]: I0313 07:57:32.391256 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7e32caff-0034-47f2-aa5b-917607dd44b6","Type":"ContainerStarted","Data":"c8f22644d11eab93b49f620b19d59b68f5003ffe111da422e203be62d357d1fe"} Mar 13 07:57:32 crc kubenswrapper[4876]: I0313 07:57:32.414202 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=27.988119627 podStartE2EDuration="38.414184761s" podCreationTimestamp="2026-03-13 07:56:54 +0000 UTC" firstStartedPulling="2026-03-13 07:57:14.780889712 +0000 UTC m=+1094.451668684" lastFinishedPulling="2026-03-13 07:57:25.206954836 +0000 UTC m=+1104.877733818" observedRunningTime="2026-03-13 07:57:32.410750053 +0000 UTC m=+1112.081529035" watchObservedRunningTime="2026-03-13 07:57:32.414184761 +0000 UTC m=+1112.084963743" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.253419 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.253543 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.305318 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.433079 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.486265 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.567645 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.767600 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-bmxln"] Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.769858 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.779522 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.788376 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-bmxln"] Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.903543 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.903771 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-config\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.904176 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4cpr\" (UniqueName: \"kubernetes.io/projected/9e2063f7-e0a7-4aac-a9fc-08fabed80869-kube-api-access-x4cpr\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.904306 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.968827 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-rgh2p"] Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.970068 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.974004 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 13 07:57:33 crc kubenswrapper[4876]: I0313 07:57:33.979624 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rgh2p"] Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.023789 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-config\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.023972 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4cpr\" (UniqueName: \"kubernetes.io/projected/9e2063f7-e0a7-4aac-a9fc-08fabed80869-kube-api-access-x4cpr\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.024006 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.024079 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.030013 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-config\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.030686 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.032490 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.049662 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4cpr\" (UniqueName: \"kubernetes.io/projected/9e2063f7-e0a7-4aac-a9fc-08fabed80869-kube-api-access-x4cpr\") pod \"dnsmasq-dns-7fd796d7df-bmxln\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.090707 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.126722 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-ovs-rundir\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.127123 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.127148 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-config\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.127179 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-ovn-rundir\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.127203 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-combined-ca-bundle\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.127222 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqd8j\" (UniqueName: \"kubernetes.io/projected/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-kube-api-access-wqd8j\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.229188 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-ovs-rundir\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.229266 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.229299 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-config\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.229354 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-ovn-rundir\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.229388 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-combined-ca-bundle\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.229431 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqd8j\" (UniqueName: \"kubernetes.io/projected/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-kube-api-access-wqd8j\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.230404 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-ovs-rundir\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.231386 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-ovn-rundir\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.233606 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-config\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.238668 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.241201 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-combined-ca-bundle\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.246628 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqd8j\" (UniqueName: \"kubernetes.io/projected/d0dbe7f9-9cce-4b19-86d6-1b445c996d95-kube-api-access-wqd8j\") pod \"ovn-controller-metrics-rgh2p\" (UID: \"d0dbe7f9-9cce-4b19-86d6-1b445c996d95\") " pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.287434 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rgh2p" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.434134 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-bmxln"] Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.480432 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-bmxln"] Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.583336 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7hrws"] Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.584992 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.594492 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.615590 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7hrws"] Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.643605 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-config\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.643661 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.643698 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.643779 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsp2h\" (UniqueName: \"kubernetes.io/projected/8fd39b9d-50d3-4a6d-8328-0d4155097d25-kube-api-access-lsp2h\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.644106 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.680603 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.745469 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsp2h\" (UniqueName: \"kubernetes.io/projected/8fd39b9d-50d3-4a6d-8328-0d4155097d25-kube-api-access-lsp2h\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.745557 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.745590 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-config\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.745628 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.745670 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.746697 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.746951 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.747217 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-config\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.757514 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.790930 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsp2h\" (UniqueName: \"kubernetes.io/projected/8fd39b9d-50d3-4a6d-8328-0d4155097d25-kube-api-access-lsp2h\") pod \"dnsmasq-dns-86db49b7ff-7hrws\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.947147 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.953707 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.958192 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.958468 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.958616 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.965800 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.968451 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-6bqql" Mar 13 07:57:34 crc kubenswrapper[4876]: I0313 07:57:34.983326 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.057283 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.057391 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq8ft\" (UniqueName: \"kubernetes.io/projected/76257d47-12f0-4e22-a3d7-856367d8a08b-kube-api-access-kq8ft\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.057445 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76257d47-12f0-4e22-a3d7-856367d8a08b-config\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.057475 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76257d47-12f0-4e22-a3d7-856367d8a08b-scripts\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.057525 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.057551 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/76257d47-12f0-4e22-a3d7-856367d8a08b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.057581 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.088535 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rgh2p"] Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.159955 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.160030 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/76257d47-12f0-4e22-a3d7-856367d8a08b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.160059 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.160754 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.160837 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq8ft\" (UniqueName: \"kubernetes.io/projected/76257d47-12f0-4e22-a3d7-856367d8a08b-kube-api-access-kq8ft\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.160940 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76257d47-12f0-4e22-a3d7-856367d8a08b-config\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.161023 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76257d47-12f0-4e22-a3d7-856367d8a08b-scripts\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.161027 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/76257d47-12f0-4e22-a3d7-856367d8a08b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.162143 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76257d47-12f0-4e22-a3d7-856367d8a08b-config\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.162684 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76257d47-12f0-4e22-a3d7-856367d8a08b-scripts\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.165823 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.166379 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.167526 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76257d47-12f0-4e22-a3d7-856367d8a08b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.184155 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq8ft\" (UniqueName: \"kubernetes.io/projected/76257d47-12f0-4e22-a3d7-856367d8a08b-kube-api-access-kq8ft\") pod \"ovn-northd-0\" (UID: \"76257d47-12f0-4e22-a3d7-856367d8a08b\") " pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.285537 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.424892 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.425479 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.427493 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rgh2p" event={"ID":"d0dbe7f9-9cce-4b19-86d6-1b445c996d95","Type":"ContainerStarted","Data":"7e57fbfe7999d1dcb3f9e431bf4aae02f8d8b2444b050ce49569946ee799bd33"} Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.432840 4876 generic.go:334] "Generic (PLEG): container finished" podID="9e2063f7-e0a7-4aac-a9fc-08fabed80869" containerID="d5a057b4e2511c825818aeb1edc01030eb92ae12010f21552ea5b666b0a052fa" exitCode=0 Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.433329 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" event={"ID":"9e2063f7-e0a7-4aac-a9fc-08fabed80869","Type":"ContainerDied","Data":"d5a057b4e2511c825818aeb1edc01030eb92ae12010f21552ea5b666b0a052fa"} Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.433422 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" event={"ID":"9e2063f7-e0a7-4aac-a9fc-08fabed80869","Type":"ContainerStarted","Data":"2aab79a9b8e5d3b84765c67b7b9358f6126bc2c8a83dd798393fbe126ec8501f"} Mar 13 07:57:35 crc kubenswrapper[4876]: W0313 07:57:35.602931 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fd39b9d_50d3_4a6d_8328_0d4155097d25.slice/crio-a3eb01cab17e22e1cb899ffc66a648c76376bef8e80b345176862e01a769b351 WatchSource:0}: Error finding container a3eb01cab17e22e1cb899ffc66a648c76376bef8e80b345176862e01a769b351: Status 404 returned error can't find the container with id a3eb01cab17e22e1cb899ffc66a648c76376bef8e80b345176862e01a769b351 Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.605009 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7hrws"] Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.778121 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.792212 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.875428 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-dns-svc\") pod \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.875512 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-ovsdbserver-nb\") pod \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.875670 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-config\") pod \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.875747 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4cpr\" (UniqueName: \"kubernetes.io/projected/9e2063f7-e0a7-4aac-a9fc-08fabed80869-kube-api-access-x4cpr\") pod \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\" (UID: \"9e2063f7-e0a7-4aac-a9fc-08fabed80869\") " Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.925968 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e2063f7-e0a7-4aac-a9fc-08fabed80869-kube-api-access-x4cpr" (OuterVolumeSpecName: "kube-api-access-x4cpr") pod "9e2063f7-e0a7-4aac-a9fc-08fabed80869" (UID: "9e2063f7-e0a7-4aac-a9fc-08fabed80869"). InnerVolumeSpecName "kube-api-access-x4cpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.927805 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e2063f7-e0a7-4aac-a9fc-08fabed80869" (UID: "9e2063f7-e0a7-4aac-a9fc-08fabed80869"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.928053 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-config" (OuterVolumeSpecName: "config") pod "9e2063f7-e0a7-4aac-a9fc-08fabed80869" (UID: "9e2063f7-e0a7-4aac-a9fc-08fabed80869"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.930017 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e2063f7-e0a7-4aac-a9fc-08fabed80869" (UID: "9e2063f7-e0a7-4aac-a9fc-08fabed80869"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.977927 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4cpr\" (UniqueName: \"kubernetes.io/projected/9e2063f7-e0a7-4aac-a9fc-08fabed80869-kube-api-access-x4cpr\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.977987 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.977997 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:35 crc kubenswrapper[4876]: I0313 07:57:35.978006 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e2063f7-e0a7-4aac-a9fc-08fabed80869-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.448757 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rgh2p" event={"ID":"d0dbe7f9-9cce-4b19-86d6-1b445c996d95","Type":"ContainerStarted","Data":"f0e15098dd0d7a230563bfbeacc316bf4cd8af893a6c2bf4f91f694d3904c2ce"} Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.452499 4876 generic.go:334] "Generic (PLEG): container finished" podID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" containerID="807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5" exitCode=0 Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.452603 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" event={"ID":"8fd39b9d-50d3-4a6d-8328-0d4155097d25","Type":"ContainerDied","Data":"807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5"} Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.452653 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" event={"ID":"8fd39b9d-50d3-4a6d-8328-0d4155097d25","Type":"ContainerStarted","Data":"a3eb01cab17e22e1cb899ffc66a648c76376bef8e80b345176862e01a769b351"} Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.455379 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" event={"ID":"9e2063f7-e0a7-4aac-a9fc-08fabed80869","Type":"ContainerDied","Data":"2aab79a9b8e5d3b84765c67b7b9358f6126bc2c8a83dd798393fbe126ec8501f"} Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.455453 4876 scope.go:117] "RemoveContainer" containerID="d5a057b4e2511c825818aeb1edc01030eb92ae12010f21552ea5b666b0a052fa" Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.455448 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-bmxln" Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.457142 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"76257d47-12f0-4e22-a3d7-856367d8a08b","Type":"ContainerStarted","Data":"53c271d58827882001ec62206f027188a210e5e222208528156e100dd635b3d7"} Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.492224 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-rgh2p" podStartSLOduration=3.49220381 podStartE2EDuration="3.49220381s" podCreationTimestamp="2026-03-13 07:57:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:57:36.489264736 +0000 UTC m=+1116.160043728" watchObservedRunningTime="2026-03-13 07:57:36.49220381 +0000 UTC m=+1116.162982792" Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.590709 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-bmxln"] Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.604184 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-bmxln"] Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.894400 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.895002 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 07:57:36 crc kubenswrapper[4876]: I0313 07:57:36.996785 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 07:57:37 crc kubenswrapper[4876]: I0313 07:57:37.045880 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e2063f7-e0a7-4aac-a9fc-08fabed80869" path="/var/lib/kubelet/pods/9e2063f7-e0a7-4aac-a9fc-08fabed80869/volumes" Mar 13 07:57:37 crc kubenswrapper[4876]: I0313 07:57:37.471276 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"76257d47-12f0-4e22-a3d7-856367d8a08b","Type":"ContainerStarted","Data":"ed11b81fccb9dd2311390536cbb1f9fbeb2e9d3d51b3cb4eec680b941d9957f0"} Mar 13 07:57:37 crc kubenswrapper[4876]: I0313 07:57:37.475903 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" event={"ID":"8fd39b9d-50d3-4a6d-8328-0d4155097d25","Type":"ContainerStarted","Data":"aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7"} Mar 13 07:57:37 crc kubenswrapper[4876]: I0313 07:57:37.477223 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:37 crc kubenswrapper[4876]: I0313 07:57:37.524826 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" podStartSLOduration=3.524804101 podStartE2EDuration="3.524804101s" podCreationTimestamp="2026-03-13 07:57:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:57:37.515945879 +0000 UTC m=+1117.186724872" watchObservedRunningTime="2026-03-13 07:57:37.524804101 +0000 UTC m=+1117.195583093" Mar 13 07:57:37 crc kubenswrapper[4876]: I0313 07:57:37.703100 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.290646 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.369131 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.492994 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"76257d47-12f0-4e22-a3d7-856367d8a08b","Type":"ContainerStarted","Data":"3c76dd62e9b86a1567b7de13099112efd8954926ded5acb8c2da3d655f2fea21"} Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.518579 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.237448366 podStartE2EDuration="4.518553249s" podCreationTimestamp="2026-03-13 07:57:34 +0000 UTC" firstStartedPulling="2026-03-13 07:57:35.805513707 +0000 UTC m=+1115.476292689" lastFinishedPulling="2026-03-13 07:57:37.08661859 +0000 UTC m=+1116.757397572" observedRunningTime="2026-03-13 07:57:38.511293053 +0000 UTC m=+1118.182072035" watchObservedRunningTime="2026-03-13 07:57:38.518553249 +0000 UTC m=+1118.189332231" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.602905 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-t8c2h"] Mar 13 07:57:38 crc kubenswrapper[4876]: E0313 07:57:38.603366 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2063f7-e0a7-4aac-a9fc-08fabed80869" containerName="init" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.603386 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2063f7-e0a7-4aac-a9fc-08fabed80869" containerName="init" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.603577 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2063f7-e0a7-4aac-a9fc-08fabed80869" containerName="init" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.604159 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.616584 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-t8c2h"] Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.709737 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-70c0-account-create-update-bvqs5"] Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.711170 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.713922 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.716812 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-70c0-account-create-update-bvqs5"] Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.746622 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61f27a21-4172-4552-a738-314cceb0acd7-operator-scripts\") pod \"placement-db-create-t8c2h\" (UID: \"61f27a21-4172-4552-a738-314cceb0acd7\") " pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.746792 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8qg7\" (UniqueName: \"kubernetes.io/projected/61f27a21-4172-4552-a738-314cceb0acd7-kube-api-access-z8qg7\") pod \"placement-db-create-t8c2h\" (UID: \"61f27a21-4172-4552-a738-314cceb0acd7\") " pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.848588 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8qg7\" (UniqueName: \"kubernetes.io/projected/61f27a21-4172-4552-a738-314cceb0acd7-kube-api-access-z8qg7\") pod \"placement-db-create-t8c2h\" (UID: \"61f27a21-4172-4552-a738-314cceb0acd7\") " pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.848708 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61f27a21-4172-4552-a738-314cceb0acd7-operator-scripts\") pod \"placement-db-create-t8c2h\" (UID: \"61f27a21-4172-4552-a738-314cceb0acd7\") " pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.848742 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0341c6d0-9f56-4dcd-a167-7152687cf2c1-operator-scripts\") pod \"placement-70c0-account-create-update-bvqs5\" (UID: \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\") " pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.848798 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46q6c\" (UniqueName: \"kubernetes.io/projected/0341c6d0-9f56-4dcd-a167-7152687cf2c1-kube-api-access-46q6c\") pod \"placement-70c0-account-create-update-bvqs5\" (UID: \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\") " pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.849755 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61f27a21-4172-4552-a738-314cceb0acd7-operator-scripts\") pod \"placement-db-create-t8c2h\" (UID: \"61f27a21-4172-4552-a738-314cceb0acd7\") " pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.866721 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8qg7\" (UniqueName: \"kubernetes.io/projected/61f27a21-4172-4552-a738-314cceb0acd7-kube-api-access-z8qg7\") pod \"placement-db-create-t8c2h\" (UID: \"61f27a21-4172-4552-a738-314cceb0acd7\") " pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.951145 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0341c6d0-9f56-4dcd-a167-7152687cf2c1-operator-scripts\") pod \"placement-70c0-account-create-update-bvqs5\" (UID: \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\") " pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.951266 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46q6c\" (UniqueName: \"kubernetes.io/projected/0341c6d0-9f56-4dcd-a167-7152687cf2c1-kube-api-access-46q6c\") pod \"placement-70c0-account-create-update-bvqs5\" (UID: \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\") " pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.951613 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.952757 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0341c6d0-9f56-4dcd-a167-7152687cf2c1-operator-scripts\") pod \"placement-70c0-account-create-update-bvqs5\" (UID: \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\") " pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:38 crc kubenswrapper[4876]: I0313 07:57:38.975864 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46q6c\" (UniqueName: \"kubernetes.io/projected/0341c6d0-9f56-4dcd-a167-7152687cf2c1-kube-api-access-46q6c\") pod \"placement-70c0-account-create-update-bvqs5\" (UID: \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\") " pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.027161 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.455801 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-t8c2h"] Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.529609 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-t8c2h" event={"ID":"61f27a21-4172-4552-a738-314cceb0acd7","Type":"ContainerStarted","Data":"abe2e1db0243ffdd0184c3aa6e9eaf4641bf1094baf7ebc48781cf01cec00fc9"} Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.530131 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.540902 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7hrws"] Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.591639 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-70c0-account-create-update-bvqs5"] Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.619106 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.624691 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-pc9zn"] Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.638551 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.665994 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pc9zn"] Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.774530 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxlk8\" (UniqueName: \"kubernetes.io/projected/7c2a1e76-01f4-4164-bdde-b7c30ff87058-kube-api-access-mxlk8\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.774595 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-config\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.774977 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.775089 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-dns-svc\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.775872 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.877960 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.878066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxlk8\" (UniqueName: \"kubernetes.io/projected/7c2a1e76-01f4-4164-bdde-b7c30ff87058-kube-api-access-mxlk8\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.878101 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-config\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.878171 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.878207 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-dns-svc\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.879455 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-dns-svc\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.880200 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-config\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.880391 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.880770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.911188 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxlk8\" (UniqueName: \"kubernetes.io/projected/7c2a1e76-01f4-4164-bdde-b7c30ff87058-kube-api-access-mxlk8\") pod \"dnsmasq-dns-698758b865-pc9zn\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:39 crc kubenswrapper[4876]: I0313 07:57:39.992634 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.537728 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pc9zn"] Mar 13 07:57:40 crc kubenswrapper[4876]: W0313 07:57:40.545152 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c2a1e76_01f4_4164_bdde_b7c30ff87058.slice/crio-91c2147e976802ffea7353a764902092d9ba57c0744581cd87415d9c68a41eee WatchSource:0}: Error finding container 91c2147e976802ffea7353a764902092d9ba57c0744581cd87415d9c68a41eee: Status 404 returned error can't find the container with id 91c2147e976802ffea7353a764902092d9ba57c0744581cd87415d9c68a41eee Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.548093 4876 generic.go:334] "Generic (PLEG): container finished" podID="61f27a21-4172-4552-a738-314cceb0acd7" containerID="c4a646bbc62a1521689dd3bef40edaa23378fa7c0632719d93ff39de8669a54c" exitCode=0 Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.548257 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-t8c2h" event={"ID":"61f27a21-4172-4552-a738-314cceb0acd7","Type":"ContainerDied","Data":"c4a646bbc62a1521689dd3bef40edaa23378fa7c0632719d93ff39de8669a54c"} Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.550453 4876 generic.go:334] "Generic (PLEG): container finished" podID="0341c6d0-9f56-4dcd-a167-7152687cf2c1" containerID="79fd805194e9fd15bbe81ceb24d961ed287f1fd670b5c8fa45a2bdb63466873e" exitCode=0 Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.550505 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-70c0-account-create-update-bvqs5" event={"ID":"0341c6d0-9f56-4dcd-a167-7152687cf2c1","Type":"ContainerDied","Data":"79fd805194e9fd15bbe81ceb24d961ed287f1fd670b5c8fa45a2bdb63466873e"} Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.550547 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-70c0-account-create-update-bvqs5" event={"ID":"0341c6d0-9f56-4dcd-a167-7152687cf2c1","Type":"ContainerStarted","Data":"c061cf662c44127304290696a7b6f9b896577e3149961eca451069be5cdcacb4"} Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.550885 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" podUID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" containerName="dnsmasq-dns" containerID="cri-o://aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7" gracePeriod=10 Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.736073 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.745684 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.750379 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mgslt" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.750719 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.750868 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.751015 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.758564 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.804479 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/77efd2bc-d283-4ab6-bd5f-b346b431ea09-cache\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.807555 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wc86\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-kube-api-access-8wc86\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.807707 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.807846 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.807961 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/77efd2bc-d283-4ab6-bd5f-b346b431ea09-lock\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.808042 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77efd2bc-d283-4ab6-bd5f-b346b431ea09-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.910608 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.910731 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.910804 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/77efd2bc-d283-4ab6-bd5f-b346b431ea09-lock\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.910861 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77efd2bc-d283-4ab6-bd5f-b346b431ea09-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.910869 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.910895 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/77efd2bc-d283-4ab6-bd5f-b346b431ea09-cache\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.911087 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wc86\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-kube-api-access-8wc86\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.912088 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/77efd2bc-d283-4ab6-bd5f-b346b431ea09-lock\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: E0313 07:57:40.912305 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 07:57:40 crc kubenswrapper[4876]: E0313 07:57:40.912350 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.912399 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/77efd2bc-d283-4ab6-bd5f-b346b431ea09-cache\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: E0313 07:57:40.912429 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift podName:77efd2bc-d283-4ab6-bd5f-b346b431ea09 nodeName:}" failed. No retries permitted until 2026-03-13 07:57:41.412405311 +0000 UTC m=+1121.083184293 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift") pod "swift-storage-0" (UID: "77efd2bc-d283-4ab6-bd5f-b346b431ea09") : configmap "swift-ring-files" not found Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.925667 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77efd2bc-d283-4ab6-bd5f-b346b431ea09-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.938210 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wc86\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-kube-api-access-8wc86\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:40 crc kubenswrapper[4876]: I0313 07:57:40.957291 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.080366 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.216132 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-dns-svc\") pod \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.216330 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-config\") pod \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.216394 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-sb\") pod \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.216454 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-nb\") pod \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.216574 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsp2h\" (UniqueName: \"kubernetes.io/projected/8fd39b9d-50d3-4a6d-8328-0d4155097d25-kube-api-access-lsp2h\") pod \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\" (UID: \"8fd39b9d-50d3-4a6d-8328-0d4155097d25\") " Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.225754 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd39b9d-50d3-4a6d-8328-0d4155097d25-kube-api-access-lsp2h" (OuterVolumeSpecName: "kube-api-access-lsp2h") pod "8fd39b9d-50d3-4a6d-8328-0d4155097d25" (UID: "8fd39b9d-50d3-4a6d-8328-0d4155097d25"). InnerVolumeSpecName "kube-api-access-lsp2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.263120 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-config" (OuterVolumeSpecName: "config") pod "8fd39b9d-50d3-4a6d-8328-0d4155097d25" (UID: "8fd39b9d-50d3-4a6d-8328-0d4155097d25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.264435 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8fd39b9d-50d3-4a6d-8328-0d4155097d25" (UID: "8fd39b9d-50d3-4a6d-8328-0d4155097d25"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.279669 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8fd39b9d-50d3-4a6d-8328-0d4155097d25" (UID: "8fd39b9d-50d3-4a6d-8328-0d4155097d25"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.284101 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8fd39b9d-50d3-4a6d-8328-0d4155097d25" (UID: "8fd39b9d-50d3-4a6d-8328-0d4155097d25"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.319252 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.319294 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.319305 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.319316 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fd39b9d-50d3-4a6d-8328-0d4155097d25-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.319329 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsp2h\" (UniqueName: \"kubernetes.io/projected/8fd39b9d-50d3-4a6d-8328-0d4155097d25-kube-api-access-lsp2h\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.421204 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:41 crc kubenswrapper[4876]: E0313 07:57:41.421493 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 07:57:41 crc kubenswrapper[4876]: E0313 07:57:41.421511 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 07:57:41 crc kubenswrapper[4876]: E0313 07:57:41.421563 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift podName:77efd2bc-d283-4ab6-bd5f-b346b431ea09 nodeName:}" failed. No retries permitted until 2026-03-13 07:57:42.421547548 +0000 UTC m=+1122.092326530 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift") pod "swift-storage-0" (UID: "77efd2bc-d283-4ab6-bd5f-b346b431ea09") : configmap "swift-ring-files" not found Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.560638 4876 generic.go:334] "Generic (PLEG): container finished" podID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" containerID="5106a266925ccb985fb512f2a4c255255633daf36f8d7f87d07779a214bc2c34" exitCode=0 Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.560753 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pc9zn" event={"ID":"7c2a1e76-01f4-4164-bdde-b7c30ff87058","Type":"ContainerDied","Data":"5106a266925ccb985fb512f2a4c255255633daf36f8d7f87d07779a214bc2c34"} Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.560827 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pc9zn" event={"ID":"7c2a1e76-01f4-4164-bdde-b7c30ff87058","Type":"ContainerStarted","Data":"91c2147e976802ffea7353a764902092d9ba57c0744581cd87415d9c68a41eee"} Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.563444 4876 generic.go:334] "Generic (PLEG): container finished" podID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" containerID="aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7" exitCode=0 Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.563529 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" event={"ID":"8fd39b9d-50d3-4a6d-8328-0d4155097d25","Type":"ContainerDied","Data":"aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7"} Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.563582 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" event={"ID":"8fd39b9d-50d3-4a6d-8328-0d4155097d25","Type":"ContainerDied","Data":"a3eb01cab17e22e1cb899ffc66a648c76376bef8e80b345176862e01a769b351"} Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.563612 4876 scope.go:117] "RemoveContainer" containerID="aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.563837 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7hrws" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.633860 4876 scope.go:117] "RemoveContainer" containerID="807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.664454 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7hrws"] Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.670949 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7hrws"] Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.671618 4876 scope.go:117] "RemoveContainer" containerID="aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7" Mar 13 07:57:41 crc kubenswrapper[4876]: E0313 07:57:41.671987 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7\": container with ID starting with aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7 not found: ID does not exist" containerID="aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.672015 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7"} err="failed to get container status \"aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7\": rpc error: code = NotFound desc = could not find container \"aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7\": container with ID starting with aa8e0dffc9a40e47593e1b960d7abe4570e0ead24f179424a593e5f6620deab7 not found: ID does not exist" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.672036 4876 scope.go:117] "RemoveContainer" containerID="807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5" Mar 13 07:57:41 crc kubenswrapper[4876]: E0313 07:57:41.672420 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5\": container with ID starting with 807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5 not found: ID does not exist" containerID="807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5" Mar 13 07:57:41 crc kubenswrapper[4876]: I0313 07:57:41.672441 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5"} err="failed to get container status \"807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5\": rpc error: code = NotFound desc = could not find container \"807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5\": container with ID starting with 807efc2808d00ba8a356af3d6db4a9e481e87e4f0477b402cac341f3833541b5 not found: ID does not exist" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.388934 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-bksxd"] Mar 13 07:57:42 crc kubenswrapper[4876]: E0313 07:57:42.391094 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" containerName="init" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.391193 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" containerName="init" Mar 13 07:57:42 crc kubenswrapper[4876]: E0313 07:57:42.391288 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" containerName="dnsmasq-dns" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.391361 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" containerName="dnsmasq-dns" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.391683 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" containerName="dnsmasq-dns" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.392494 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bksxd" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.418060 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bksxd"] Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.449514 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-operator-scripts\") pod \"glance-db-create-bksxd\" (UID: \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\") " pod="openstack/glance-db-create-bksxd" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.449592 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfhkh\" (UniqueName: \"kubernetes.io/projected/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-kube-api-access-nfhkh\") pod \"glance-db-create-bksxd\" (UID: \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\") " pod="openstack/glance-db-create-bksxd" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.449633 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:42 crc kubenswrapper[4876]: E0313 07:57:42.449858 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 07:57:42 crc kubenswrapper[4876]: E0313 07:57:42.449875 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 07:57:42 crc kubenswrapper[4876]: E0313 07:57:42.449934 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift podName:77efd2bc-d283-4ab6-bd5f-b346b431ea09 nodeName:}" failed. No retries permitted until 2026-03-13 07:57:44.44991503 +0000 UTC m=+1124.120694022 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift") pod "swift-storage-0" (UID: "77efd2bc-d283-4ab6-bd5f-b346b431ea09") : configmap "swift-ring-files" not found Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.551516 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-operator-scripts\") pod \"glance-db-create-bksxd\" (UID: \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\") " pod="openstack/glance-db-create-bksxd" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.551943 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfhkh\" (UniqueName: \"kubernetes.io/projected/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-kube-api-access-nfhkh\") pod \"glance-db-create-bksxd\" (UID: \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\") " pod="openstack/glance-db-create-bksxd" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.553175 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-operator-scripts\") pod \"glance-db-create-bksxd\" (UID: \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\") " pod="openstack/glance-db-create-bksxd" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.570675 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-d7a7-account-create-update-q8h2f"] Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.573797 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.582831 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.587353 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfhkh\" (UniqueName: \"kubernetes.io/projected/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-kube-api-access-nfhkh\") pod \"glance-db-create-bksxd\" (UID: \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\") " pod="openstack/glance-db-create-bksxd" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.599856 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d7a7-account-create-update-q8h2f"] Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.657396 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4188868-e4dc-4960-a300-ebc151e4408f-operator-scripts\") pod \"glance-d7a7-account-create-update-q8h2f\" (UID: \"f4188868-e4dc-4960-a300-ebc151e4408f\") " pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.657673 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj6gg\" (UniqueName: \"kubernetes.io/projected/f4188868-e4dc-4960-a300-ebc151e4408f-kube-api-access-xj6gg\") pod \"glance-d7a7-account-create-update-q8h2f\" (UID: \"f4188868-e4dc-4960-a300-ebc151e4408f\") " pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.711025 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bksxd" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.759529 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4188868-e4dc-4960-a300-ebc151e4408f-operator-scripts\") pod \"glance-d7a7-account-create-update-q8h2f\" (UID: \"f4188868-e4dc-4960-a300-ebc151e4408f\") " pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.759610 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj6gg\" (UniqueName: \"kubernetes.io/projected/f4188868-e4dc-4960-a300-ebc151e4408f-kube-api-access-xj6gg\") pod \"glance-d7a7-account-create-update-q8h2f\" (UID: \"f4188868-e4dc-4960-a300-ebc151e4408f\") " pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.760604 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4188868-e4dc-4960-a300-ebc151e4408f-operator-scripts\") pod \"glance-d7a7-account-create-update-q8h2f\" (UID: \"f4188868-e4dc-4960-a300-ebc151e4408f\") " pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.785062 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj6gg\" (UniqueName: \"kubernetes.io/projected/f4188868-e4dc-4960-a300-ebc151e4408f-kube-api-access-xj6gg\") pod \"glance-d7a7-account-create-update-q8h2f\" (UID: \"f4188868-e4dc-4960-a300-ebc151e4408f\") " pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:42 crc kubenswrapper[4876]: I0313 07:57:42.943736 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:43 crc kubenswrapper[4876]: I0313 07:57:43.045843 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd39b9d-50d3-4a6d-8328-0d4155097d25" path="/var/lib/kubelet/pods/8fd39b9d-50d3-4a6d-8328-0d4155097d25/volumes" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.039316 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-gpnkg"] Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.040830 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.043734 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.054403 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gpnkg"] Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.084896 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmsg2\" (UniqueName: \"kubernetes.io/projected/3ba6da58-adca-458c-b3e8-004c8be5c1d6-kube-api-access-jmsg2\") pod \"root-account-create-update-gpnkg\" (UID: \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\") " pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.084987 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ba6da58-adca-458c-b3e8-004c8be5c1d6-operator-scripts\") pod \"root-account-create-update-gpnkg\" (UID: \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\") " pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.187542 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmsg2\" (UniqueName: \"kubernetes.io/projected/3ba6da58-adca-458c-b3e8-004c8be5c1d6-kube-api-access-jmsg2\") pod \"root-account-create-update-gpnkg\" (UID: \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\") " pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.187646 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ba6da58-adca-458c-b3e8-004c8be5c1d6-operator-scripts\") pod \"root-account-create-update-gpnkg\" (UID: \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\") " pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.188413 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ba6da58-adca-458c-b3e8-004c8be5c1d6-operator-scripts\") pod \"root-account-create-update-gpnkg\" (UID: \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\") " pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.213022 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmsg2\" (UniqueName: \"kubernetes.io/projected/3ba6da58-adca-458c-b3e8-004c8be5c1d6-kube-api-access-jmsg2\") pod \"root-account-create-update-gpnkg\" (UID: \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\") " pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.362977 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.493459 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:44 crc kubenswrapper[4876]: E0313 07:57:44.493730 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 07:57:44 crc kubenswrapper[4876]: E0313 07:57:44.493767 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 07:57:44 crc kubenswrapper[4876]: E0313 07:57:44.493846 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift podName:77efd2bc-d283-4ab6-bd5f-b346b431ea09 nodeName:}" failed. No retries permitted until 2026-03-13 07:57:48.493825028 +0000 UTC m=+1128.164604000 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift") pod "swift-storage-0" (UID: "77efd2bc-d283-4ab6-bd5f-b346b431ea09") : configmap "swift-ring-files" not found Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.664402 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-pbnh2"] Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.665560 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.667927 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.669549 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.677752 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.679471 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-pbnh2"] Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.802639 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5294658-a1f0-4340-9d61-c8008a37b5e9-etc-swift\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.802734 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-swiftconf\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.802778 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59ths\" (UniqueName: \"kubernetes.io/projected/e5294658-a1f0-4340-9d61-c8008a37b5e9-kube-api-access-59ths\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.802919 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-dispersionconf\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.803047 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-ring-data-devices\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.803286 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-scripts\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.803310 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-combined-ca-bundle\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.905397 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-dispersionconf\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.905489 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-ring-data-devices\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.905594 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-scripts\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.905665 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-combined-ca-bundle\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.906537 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-ring-data-devices\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.906841 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-scripts\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.907041 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5294658-a1f0-4340-9d61-c8008a37b5e9-etc-swift\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.907115 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-swiftconf\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.907697 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59ths\" (UniqueName: \"kubernetes.io/projected/e5294658-a1f0-4340-9d61-c8008a37b5e9-kube-api-access-59ths\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.907460 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5294658-a1f0-4340-9d61-c8008a37b5e9-etc-swift\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.909335 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-combined-ca-bundle\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.910024 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-swiftconf\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.915111 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-dispersionconf\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.933643 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59ths\" (UniqueName: \"kubernetes.io/projected/e5294658-a1f0-4340-9d61-c8008a37b5e9-kube-api-access-59ths\") pod \"swift-ring-rebalance-pbnh2\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:44 crc kubenswrapper[4876]: I0313 07:57:44.989068 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.534901 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.560630 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.621826 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0341c6d0-9f56-4dcd-a167-7152687cf2c1-operator-scripts\") pod \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\" (UID: \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\") " Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.621887 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61f27a21-4172-4552-a738-314cceb0acd7-operator-scripts\") pod \"61f27a21-4172-4552-a738-314cceb0acd7\" (UID: \"61f27a21-4172-4552-a738-314cceb0acd7\") " Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.621967 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46q6c\" (UniqueName: \"kubernetes.io/projected/0341c6d0-9f56-4dcd-a167-7152687cf2c1-kube-api-access-46q6c\") pod \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\" (UID: \"0341c6d0-9f56-4dcd-a167-7152687cf2c1\") " Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.621995 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8qg7\" (UniqueName: \"kubernetes.io/projected/61f27a21-4172-4552-a738-314cceb0acd7-kube-api-access-z8qg7\") pod \"61f27a21-4172-4552-a738-314cceb0acd7\" (UID: \"61f27a21-4172-4552-a738-314cceb0acd7\") " Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.635296 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f27a21-4172-4552-a738-314cceb0acd7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "61f27a21-4172-4552-a738-314cceb0acd7" (UID: "61f27a21-4172-4552-a738-314cceb0acd7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.635743 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0341c6d0-9f56-4dcd-a167-7152687cf2c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0341c6d0-9f56-4dcd-a167-7152687cf2c1" (UID: "0341c6d0-9f56-4dcd-a167-7152687cf2c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.638839 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f27a21-4172-4552-a738-314cceb0acd7-kube-api-access-z8qg7" (OuterVolumeSpecName: "kube-api-access-z8qg7") pod "61f27a21-4172-4552-a738-314cceb0acd7" (UID: "61f27a21-4172-4552-a738-314cceb0acd7"). InnerVolumeSpecName "kube-api-access-z8qg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.647307 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0341c6d0-9f56-4dcd-a167-7152687cf2c1-kube-api-access-46q6c" (OuterVolumeSpecName: "kube-api-access-46q6c") pod "0341c6d0-9f56-4dcd-a167-7152687cf2c1" (UID: "0341c6d0-9f56-4dcd-a167-7152687cf2c1"). InnerVolumeSpecName "kube-api-access-46q6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.666609 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-70c0-account-create-update-bvqs5" event={"ID":"0341c6d0-9f56-4dcd-a167-7152687cf2c1","Type":"ContainerDied","Data":"c061cf662c44127304290696a7b6f9b896577e3149961eca451069be5cdcacb4"} Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.666663 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c061cf662c44127304290696a7b6f9b896577e3149961eca451069be5cdcacb4" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.666746 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-70c0-account-create-update-bvqs5" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.700808 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-t8c2h" event={"ID":"61f27a21-4172-4552-a738-314cceb0acd7","Type":"ContainerDied","Data":"abe2e1db0243ffdd0184c3aa6e9eaf4641bf1094baf7ebc48781cf01cec00fc9"} Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.700883 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abe2e1db0243ffdd0184c3aa6e9eaf4641bf1094baf7ebc48781cf01cec00fc9" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.700966 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-t8c2h" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.723914 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0341c6d0-9f56-4dcd-a167-7152687cf2c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.723949 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61f27a21-4172-4552-a738-314cceb0acd7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.723961 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46q6c\" (UniqueName: \"kubernetes.io/projected/0341c6d0-9f56-4dcd-a167-7152687cf2c1-kube-api-access-46q6c\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.723971 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8qg7\" (UniqueName: \"kubernetes.io/projected/61f27a21-4172-4552-a738-314cceb0acd7-kube-api-access-z8qg7\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:45 crc kubenswrapper[4876]: I0313 07:57:45.990208 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bksxd"] Mar 13 07:57:46 crc kubenswrapper[4876]: I0313 07:57:46.216297 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d7a7-account-create-update-q8h2f"] Mar 13 07:57:46 crc kubenswrapper[4876]: W0313 07:57:46.218888 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4188868_e4dc_4960_a300_ebc151e4408f.slice/crio-92e04327a4a779f31deb632cc81a89c383288d6669ed48e7640ea19c10175e00 WatchSource:0}: Error finding container 92e04327a4a779f31deb632cc81a89c383288d6669ed48e7640ea19c10175e00: Status 404 returned error can't find the container with id 92e04327a4a779f31deb632cc81a89c383288d6669ed48e7640ea19c10175e00 Mar 13 07:57:46 crc kubenswrapper[4876]: I0313 07:57:46.235325 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-pbnh2"] Mar 13 07:57:46 crc kubenswrapper[4876]: W0313 07:57:46.270664 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ba6da58_adca_458c_b3e8_004c8be5c1d6.slice/crio-d3efb7cbb85cfc7b6dd9eeac9339f176a11c20c0e4f4ad58d9af20225678620a WatchSource:0}: Error finding container d3efb7cbb85cfc7b6dd9eeac9339f176a11c20c0e4f4ad58d9af20225678620a: Status 404 returned error can't find the container with id d3efb7cbb85cfc7b6dd9eeac9339f176a11c20c0e4f4ad58d9af20225678620a Mar 13 07:57:46 crc kubenswrapper[4876]: I0313 07:57:46.272730 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gpnkg"] Mar 13 07:57:46 crc kubenswrapper[4876]: I0313 07:57:46.709018 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gpnkg" event={"ID":"3ba6da58-adca-458c-b3e8-004c8be5c1d6","Type":"ContainerStarted","Data":"d3efb7cbb85cfc7b6dd9eeac9339f176a11c20c0e4f4ad58d9af20225678620a"} Mar 13 07:57:46 crc kubenswrapper[4876]: I0313 07:57:46.715549 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pbnh2" event={"ID":"e5294658-a1f0-4340-9d61-c8008a37b5e9","Type":"ContainerStarted","Data":"104b1d585cfafe6c568d3d9bf9d008b9540d8fc06483dd249832520dbeff227d"} Mar 13 07:57:46 crc kubenswrapper[4876]: I0313 07:57:46.716718 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bksxd" event={"ID":"891ee27b-6971-41e8-baee-ea2f4b4a0a4e","Type":"ContainerStarted","Data":"02c180f4b0afa57aa56ff0a05db3b68c2af775f1a12e89fce5694cdc72b7551a"} Mar 13 07:57:46 crc kubenswrapper[4876]: I0313 07:57:46.717887 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d7a7-account-create-update-q8h2f" event={"ID":"f4188868-e4dc-4960-a300-ebc151e4408f","Type":"ContainerStarted","Data":"92e04327a4a779f31deb632cc81a89c383288d6669ed48e7640ea19c10175e00"} Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.729095 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bksxd" event={"ID":"891ee27b-6971-41e8-baee-ea2f4b4a0a4e","Type":"ContainerStarted","Data":"a8f830fcaa939cb42366cb37d1e9ebaed6171d66f16a8e1d4ab81fa1711c3eda"} Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.734722 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d7a7-account-create-update-q8h2f" event={"ID":"f4188868-e4dc-4960-a300-ebc151e4408f","Type":"ContainerStarted","Data":"f922a95faaaf45391c12a6d61c2a2671af6fa3251036f30d410a0d7a480ad2e4"} Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.737558 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pc9zn" event={"ID":"7c2a1e76-01f4-4164-bdde-b7c30ff87058","Type":"ContainerStarted","Data":"63b8adbf89f8a0d05bacf662084557cbc5e86c6409efe944b5ea1eaec49f08e2"} Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.738265 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.739879 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gpnkg" event={"ID":"3ba6da58-adca-458c-b3e8-004c8be5c1d6","Type":"ContainerStarted","Data":"d7e72f20ba8a3569864c6741c938b22f3bb790db5a73b9bb43f7e825d3a9aa14"} Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.766211 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-bksxd" podStartSLOduration=5.766189083 podStartE2EDuration="5.766189083s" podCreationTimestamp="2026-03-13 07:57:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:57:47.757421263 +0000 UTC m=+1127.428200255" watchObservedRunningTime="2026-03-13 07:57:47.766189083 +0000 UTC m=+1127.436968065" Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.786214 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-pc9zn" podStartSLOduration=8.786185021 podStartE2EDuration="8.786185021s" podCreationTimestamp="2026-03-13 07:57:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:57:47.783320209 +0000 UTC m=+1127.454099191" watchObservedRunningTime="2026-03-13 07:57:47.786185021 +0000 UTC m=+1127.456964003" Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.803437 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-d7a7-account-create-update-q8h2f" podStartSLOduration=5.80341391 podStartE2EDuration="5.80341391s" podCreationTimestamp="2026-03-13 07:57:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:57:47.798269035 +0000 UTC m=+1127.469048027" watchObservedRunningTime="2026-03-13 07:57:47.80341391 +0000 UTC m=+1127.474192892" Mar 13 07:57:47 crc kubenswrapper[4876]: I0313 07:57:47.820948 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-gpnkg" podStartSLOduration=3.820924098 podStartE2EDuration="3.820924098s" podCreationTimestamp="2026-03-13 07:57:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:57:47.81358743 +0000 UTC m=+1127.484366412" watchObservedRunningTime="2026-03-13 07:57:47.820924098 +0000 UTC m=+1127.491703090" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.270772 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jkf72"] Mar 13 07:57:48 crc kubenswrapper[4876]: E0313 07:57:48.271197 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0341c6d0-9f56-4dcd-a167-7152687cf2c1" containerName="mariadb-account-create-update" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.271216 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0341c6d0-9f56-4dcd-a167-7152687cf2c1" containerName="mariadb-account-create-update" Mar 13 07:57:48 crc kubenswrapper[4876]: E0313 07:57:48.271260 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f27a21-4172-4552-a738-314cceb0acd7" containerName="mariadb-database-create" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.271266 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f27a21-4172-4552-a738-314cceb0acd7" containerName="mariadb-database-create" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.271435 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f27a21-4172-4552-a738-314cceb0acd7" containerName="mariadb-database-create" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.271460 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0341c6d0-9f56-4dcd-a167-7152687cf2c1" containerName="mariadb-account-create-update" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.272017 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.299905 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jkf72"] Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.381776 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13197959-749d-4222-ae6c-36fe74c289ff-operator-scripts\") pod \"keystone-db-create-jkf72\" (UID: \"13197959-749d-4222-ae6c-36fe74c289ff\") " pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.382208 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbbpc\" (UniqueName: \"kubernetes.io/projected/13197959-749d-4222-ae6c-36fe74c289ff-kube-api-access-wbbpc\") pod \"keystone-db-create-jkf72\" (UID: \"13197959-749d-4222-ae6c-36fe74c289ff\") " pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.386400 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8359-account-create-update-fr5tj"] Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.388185 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.391004 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.394914 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8359-account-create-update-fr5tj"] Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.484017 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13197959-749d-4222-ae6c-36fe74c289ff-operator-scripts\") pod \"keystone-db-create-jkf72\" (UID: \"13197959-749d-4222-ae6c-36fe74c289ff\") " pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.484101 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32838d97-317e-4102-9fdc-d707329a4f45-operator-scripts\") pod \"keystone-8359-account-create-update-fr5tj\" (UID: \"32838d97-317e-4102-9fdc-d707329a4f45\") " pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.484138 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbbpc\" (UniqueName: \"kubernetes.io/projected/13197959-749d-4222-ae6c-36fe74c289ff-kube-api-access-wbbpc\") pod \"keystone-db-create-jkf72\" (UID: \"13197959-749d-4222-ae6c-36fe74c289ff\") " pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.484221 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl75p\" (UniqueName: \"kubernetes.io/projected/32838d97-317e-4102-9fdc-d707329a4f45-kube-api-access-kl75p\") pod \"keystone-8359-account-create-update-fr5tj\" (UID: \"32838d97-317e-4102-9fdc-d707329a4f45\") " pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.485511 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13197959-749d-4222-ae6c-36fe74c289ff-operator-scripts\") pod \"keystone-db-create-jkf72\" (UID: \"13197959-749d-4222-ae6c-36fe74c289ff\") " pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.506144 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbbpc\" (UniqueName: \"kubernetes.io/projected/13197959-749d-4222-ae6c-36fe74c289ff-kube-api-access-wbbpc\") pod \"keystone-db-create-jkf72\" (UID: \"13197959-749d-4222-ae6c-36fe74c289ff\") " pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.585567 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32838d97-317e-4102-9fdc-d707329a4f45-operator-scripts\") pod \"keystone-8359-account-create-update-fr5tj\" (UID: \"32838d97-317e-4102-9fdc-d707329a4f45\") " pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.585652 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.585703 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl75p\" (UniqueName: \"kubernetes.io/projected/32838d97-317e-4102-9fdc-d707329a4f45-kube-api-access-kl75p\") pod \"keystone-8359-account-create-update-fr5tj\" (UID: \"32838d97-317e-4102-9fdc-d707329a4f45\") " pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:48 crc kubenswrapper[4876]: E0313 07:57:48.586184 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 07:57:48 crc kubenswrapper[4876]: E0313 07:57:48.586206 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 07:57:48 crc kubenswrapper[4876]: E0313 07:57:48.586258 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift podName:77efd2bc-d283-4ab6-bd5f-b346b431ea09 nodeName:}" failed. No retries permitted until 2026-03-13 07:57:56.586230224 +0000 UTC m=+1136.257009206 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift") pod "swift-storage-0" (UID: "77efd2bc-d283-4ab6-bd5f-b346b431ea09") : configmap "swift-ring-files" not found Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.586576 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32838d97-317e-4102-9fdc-d707329a4f45-operator-scripts\") pod \"keystone-8359-account-create-update-fr5tj\" (UID: \"32838d97-317e-4102-9fdc-d707329a4f45\") " pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.610880 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.629302 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl75p\" (UniqueName: \"kubernetes.io/projected/32838d97-317e-4102-9fdc-d707329a4f45-kube-api-access-kl75p\") pod \"keystone-8359-account-create-update-fr5tj\" (UID: \"32838d97-317e-4102-9fdc-d707329a4f45\") " pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.709421 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.751575 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4188868-e4dc-4960-a300-ebc151e4408f" containerID="f922a95faaaf45391c12a6d61c2a2671af6fa3251036f30d410a0d7a480ad2e4" exitCode=0 Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.751635 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d7a7-account-create-update-q8h2f" event={"ID":"f4188868-e4dc-4960-a300-ebc151e4408f","Type":"ContainerDied","Data":"f922a95faaaf45391c12a6d61c2a2671af6fa3251036f30d410a0d7a480ad2e4"} Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.756614 4876 generic.go:334] "Generic (PLEG): container finished" podID="3ba6da58-adca-458c-b3e8-004c8be5c1d6" containerID="d7e72f20ba8a3569864c6741c938b22f3bb790db5a73b9bb43f7e825d3a9aa14" exitCode=0 Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.756772 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gpnkg" event={"ID":"3ba6da58-adca-458c-b3e8-004c8be5c1d6","Type":"ContainerDied","Data":"d7e72f20ba8a3569864c6741c938b22f3bb790db5a73b9bb43f7e825d3a9aa14"} Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.763718 4876 generic.go:334] "Generic (PLEG): container finished" podID="891ee27b-6971-41e8-baee-ea2f4b4a0a4e" containerID="a8f830fcaa939cb42366cb37d1e9ebaed6171d66f16a8e1d4ab81fa1711c3eda" exitCode=0 Mar 13 07:57:48 crc kubenswrapper[4876]: I0313 07:57:48.764335 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bksxd" event={"ID":"891ee27b-6971-41e8-baee-ea2f4b4a0a4e","Type":"ContainerDied","Data":"a8f830fcaa939cb42366cb37d1e9ebaed6171d66f16a8e1d4ab81fa1711c3eda"} Mar 13 07:57:50 crc kubenswrapper[4876]: I0313 07:57:50.783694 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gpnkg" event={"ID":"3ba6da58-adca-458c-b3e8-004c8be5c1d6","Type":"ContainerDied","Data":"d3efb7cbb85cfc7b6dd9eeac9339f176a11c20c0e4f4ad58d9af20225678620a"} Mar 13 07:57:50 crc kubenswrapper[4876]: I0313 07:57:50.784166 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3efb7cbb85cfc7b6dd9eeac9339f176a11c20c0e4f4ad58d9af20225678620a" Mar 13 07:57:50 crc kubenswrapper[4876]: I0313 07:57:50.785892 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bksxd" event={"ID":"891ee27b-6971-41e8-baee-ea2f4b4a0a4e","Type":"ContainerDied","Data":"02c180f4b0afa57aa56ff0a05db3b68c2af775f1a12e89fce5694cdc72b7551a"} Mar 13 07:57:50 crc kubenswrapper[4876]: I0313 07:57:50.785940 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02c180f4b0afa57aa56ff0a05db3b68c2af775f1a12e89fce5694cdc72b7551a" Mar 13 07:57:50 crc kubenswrapper[4876]: I0313 07:57:50.788742 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d7a7-account-create-update-q8h2f" event={"ID":"f4188868-e4dc-4960-a300-ebc151e4408f","Type":"ContainerDied","Data":"92e04327a4a779f31deb632cc81a89c383288d6669ed48e7640ea19c10175e00"} Mar 13 07:57:50 crc kubenswrapper[4876]: I0313 07:57:50.788771 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92e04327a4a779f31deb632cc81a89c383288d6669ed48e7640ea19c10175e00" Mar 13 07:57:50 crc kubenswrapper[4876]: I0313 07:57:50.970399 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.011195 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.034927 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bksxd" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.043308 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4188868-e4dc-4960-a300-ebc151e4408f-operator-scripts\") pod \"f4188868-e4dc-4960-a300-ebc151e4408f\" (UID: \"f4188868-e4dc-4960-a300-ebc151e4408f\") " Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.043568 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj6gg\" (UniqueName: \"kubernetes.io/projected/f4188868-e4dc-4960-a300-ebc151e4408f-kube-api-access-xj6gg\") pod \"f4188868-e4dc-4960-a300-ebc151e4408f\" (UID: \"f4188868-e4dc-4960-a300-ebc151e4408f\") " Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.046214 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4188868-e4dc-4960-a300-ebc151e4408f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f4188868-e4dc-4960-a300-ebc151e4408f" (UID: "f4188868-e4dc-4960-a300-ebc151e4408f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.050425 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4188868-e4dc-4960-a300-ebc151e4408f-kube-api-access-xj6gg" (OuterVolumeSpecName: "kube-api-access-xj6gg") pod "f4188868-e4dc-4960-a300-ebc151e4408f" (UID: "f4188868-e4dc-4960-a300-ebc151e4408f"). InnerVolumeSpecName "kube-api-access-xj6gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.145598 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-operator-scripts\") pod \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\" (UID: \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\") " Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.145662 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmsg2\" (UniqueName: \"kubernetes.io/projected/3ba6da58-adca-458c-b3e8-004c8be5c1d6-kube-api-access-jmsg2\") pod \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\" (UID: \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\") " Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.145826 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ba6da58-adca-458c-b3e8-004c8be5c1d6-operator-scripts\") pod \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\" (UID: \"3ba6da58-adca-458c-b3e8-004c8be5c1d6\") " Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.145925 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfhkh\" (UniqueName: \"kubernetes.io/projected/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-kube-api-access-nfhkh\") pod \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\" (UID: \"891ee27b-6971-41e8-baee-ea2f4b4a0a4e\") " Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.146498 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba6da58-adca-458c-b3e8-004c8be5c1d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3ba6da58-adca-458c-b3e8-004c8be5c1d6" (UID: "3ba6da58-adca-458c-b3e8-004c8be5c1d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.146748 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "891ee27b-6971-41e8-baee-ea2f4b4a0a4e" (UID: "891ee27b-6971-41e8-baee-ea2f4b4a0a4e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.147761 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj6gg\" (UniqueName: \"kubernetes.io/projected/f4188868-e4dc-4960-a300-ebc151e4408f-kube-api-access-xj6gg\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.147956 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ba6da58-adca-458c-b3e8-004c8be5c1d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.148005 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4188868-e4dc-4960-a300-ebc151e4408f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.148022 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.152522 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-kube-api-access-nfhkh" (OuterVolumeSpecName: "kube-api-access-nfhkh") pod "891ee27b-6971-41e8-baee-ea2f4b4a0a4e" (UID: "891ee27b-6971-41e8-baee-ea2f4b4a0a4e"). InnerVolumeSpecName "kube-api-access-nfhkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.152820 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ba6da58-adca-458c-b3e8-004c8be5c1d6-kube-api-access-jmsg2" (OuterVolumeSpecName: "kube-api-access-jmsg2") pod "3ba6da58-adca-458c-b3e8-004c8be5c1d6" (UID: "3ba6da58-adca-458c-b3e8-004c8be5c1d6"). InnerVolumeSpecName "kube-api-access-jmsg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.252690 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmsg2\" (UniqueName: \"kubernetes.io/projected/3ba6da58-adca-458c-b3e8-004c8be5c1d6-kube-api-access-jmsg2\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.252744 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfhkh\" (UniqueName: \"kubernetes.io/projected/891ee27b-6971-41e8-baee-ea2f4b4a0a4e-kube-api-access-nfhkh\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.311303 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8359-account-create-update-fr5tj"] Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.336181 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jkf72"] Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.798560 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pbnh2" event={"ID":"e5294658-a1f0-4340-9d61-c8008a37b5e9","Type":"ContainerStarted","Data":"3b97419c6aa9b29927297319080c64c8ec7e9e8401d46dfc9341853911354a80"} Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.801089 4876 generic.go:334] "Generic (PLEG): container finished" podID="13197959-749d-4222-ae6c-36fe74c289ff" containerID="22cf8295b2ad80a74b8ac07c17b19ea2e5caeb2f8fde49408278ee1dd60f0295" exitCode=0 Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.801148 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jkf72" event={"ID":"13197959-749d-4222-ae6c-36fe74c289ff","Type":"ContainerDied","Data":"22cf8295b2ad80a74b8ac07c17b19ea2e5caeb2f8fde49408278ee1dd60f0295"} Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.801188 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jkf72" event={"ID":"13197959-749d-4222-ae6c-36fe74c289ff","Type":"ContainerStarted","Data":"a49fe40a55dd8de11a7a5ac0b58d5d81368ca46d5e57d56e37265950d3700f0f"} Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.802784 4876 generic.go:334] "Generic (PLEG): container finished" podID="32838d97-317e-4102-9fdc-d707329a4f45" containerID="b950aa6b5a6a0e476622df730068e6056cc98e7ae5ec28122afd0b1ee7cc581e" exitCode=0 Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.802873 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d7a7-account-create-update-q8h2f" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.803142 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8359-account-create-update-fr5tj" event={"ID":"32838d97-317e-4102-9fdc-d707329a4f45","Type":"ContainerDied","Data":"b950aa6b5a6a0e476622df730068e6056cc98e7ae5ec28122afd0b1ee7cc581e"} Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.803195 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gpnkg" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.803231 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8359-account-create-update-fr5tj" event={"ID":"32838d97-317e-4102-9fdc-d707329a4f45","Type":"ContainerStarted","Data":"9262dd99a449effcbcaa291d8ce9fb5f93a46ad283f7bcae42c6b13718eaab86"} Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.803284 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bksxd" Mar 13 07:57:51 crc kubenswrapper[4876]: I0313 07:57:51.826534 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-pbnh2" podStartSLOduration=3.3008298 podStartE2EDuration="7.826509908s" podCreationTimestamp="2026-03-13 07:57:44 +0000 UTC" firstStartedPulling="2026-03-13 07:57:46.257309577 +0000 UTC m=+1125.928088559" lastFinishedPulling="2026-03-13 07:57:50.782989695 +0000 UTC m=+1130.453768667" observedRunningTime="2026-03-13 07:57:51.823952415 +0000 UTC m=+1131.494731397" watchObservedRunningTime="2026-03-13 07:57:51.826509908 +0000 UTC m=+1131.497288900" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.677120 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-ccnjk"] Mar 13 07:57:52 crc kubenswrapper[4876]: E0313 07:57:52.687055 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="891ee27b-6971-41e8-baee-ea2f4b4a0a4e" containerName="mariadb-database-create" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.687104 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="891ee27b-6971-41e8-baee-ea2f4b4a0a4e" containerName="mariadb-database-create" Mar 13 07:57:52 crc kubenswrapper[4876]: E0313 07:57:52.687120 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4188868-e4dc-4960-a300-ebc151e4408f" containerName="mariadb-account-create-update" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.687129 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4188868-e4dc-4960-a300-ebc151e4408f" containerName="mariadb-account-create-update" Mar 13 07:57:52 crc kubenswrapper[4876]: E0313 07:57:52.687165 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba6da58-adca-458c-b3e8-004c8be5c1d6" containerName="mariadb-account-create-update" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.687174 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba6da58-adca-458c-b3e8-004c8be5c1d6" containerName="mariadb-account-create-update" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.687500 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ba6da58-adca-458c-b3e8-004c8be5c1d6" containerName="mariadb-account-create-update" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.687514 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="891ee27b-6971-41e8-baee-ea2f4b4a0a4e" containerName="mariadb-database-create" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.687527 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4188868-e4dc-4960-a300-ebc151e4408f" containerName="mariadb-account-create-update" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.692841 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.695298 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.696098 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xxnmv" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.701519 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ccnjk"] Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.779735 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-db-sync-config-data\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.779783 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhvvl\" (UniqueName: \"kubernetes.io/projected/275d5189-c7c8-44f4-b753-9590b692900c-kube-api-access-bhvvl\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.779830 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-combined-ca-bundle\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.779898 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-config-data\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.882407 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-db-sync-config-data\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.882463 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhvvl\" (UniqueName: \"kubernetes.io/projected/275d5189-c7c8-44f4-b753-9590b692900c-kube-api-access-bhvvl\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.882504 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-combined-ca-bundle\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.882565 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-config-data\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.901093 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-config-data\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.901142 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-db-sync-config-data\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.906646 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-combined-ca-bundle\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:52 crc kubenswrapper[4876]: I0313 07:57:52.916292 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhvvl\" (UniqueName: \"kubernetes.io/projected/275d5189-c7c8-44f4-b753-9590b692900c-kube-api-access-bhvvl\") pod \"glance-db-sync-ccnjk\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.025638 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ccnjk" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.229461 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.267119 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.298311 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32838d97-317e-4102-9fdc-d707329a4f45-operator-scripts\") pod \"32838d97-317e-4102-9fdc-d707329a4f45\" (UID: \"32838d97-317e-4102-9fdc-d707329a4f45\") " Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.298458 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl75p\" (UniqueName: \"kubernetes.io/projected/32838d97-317e-4102-9fdc-d707329a4f45-kube-api-access-kl75p\") pod \"32838d97-317e-4102-9fdc-d707329a4f45\" (UID: \"32838d97-317e-4102-9fdc-d707329a4f45\") " Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.299222 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32838d97-317e-4102-9fdc-d707329a4f45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32838d97-317e-4102-9fdc-d707329a4f45" (UID: "32838d97-317e-4102-9fdc-d707329a4f45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.299853 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32838d97-317e-4102-9fdc-d707329a4f45-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.314183 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32838d97-317e-4102-9fdc-d707329a4f45-kube-api-access-kl75p" (OuterVolumeSpecName: "kube-api-access-kl75p") pod "32838d97-317e-4102-9fdc-d707329a4f45" (UID: "32838d97-317e-4102-9fdc-d707329a4f45"). InnerVolumeSpecName "kube-api-access-kl75p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.400949 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13197959-749d-4222-ae6c-36fe74c289ff-operator-scripts\") pod \"13197959-749d-4222-ae6c-36fe74c289ff\" (UID: \"13197959-749d-4222-ae6c-36fe74c289ff\") " Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.401031 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbbpc\" (UniqueName: \"kubernetes.io/projected/13197959-749d-4222-ae6c-36fe74c289ff-kube-api-access-wbbpc\") pod \"13197959-749d-4222-ae6c-36fe74c289ff\" (UID: \"13197959-749d-4222-ae6c-36fe74c289ff\") " Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.401569 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl75p\" (UniqueName: \"kubernetes.io/projected/32838d97-317e-4102-9fdc-d707329a4f45-kube-api-access-kl75p\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.402381 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13197959-749d-4222-ae6c-36fe74c289ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "13197959-749d-4222-ae6c-36fe74c289ff" (UID: "13197959-749d-4222-ae6c-36fe74c289ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.404599 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13197959-749d-4222-ae6c-36fe74c289ff-kube-api-access-wbbpc" (OuterVolumeSpecName: "kube-api-access-wbbpc") pod "13197959-749d-4222-ae6c-36fe74c289ff" (UID: "13197959-749d-4222-ae6c-36fe74c289ff"). InnerVolumeSpecName "kube-api-access-wbbpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.504306 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13197959-749d-4222-ae6c-36fe74c289ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.504698 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbbpc\" (UniqueName: \"kubernetes.io/projected/13197959-749d-4222-ae6c-36fe74c289ff-kube-api-access-wbbpc\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.741158 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ccnjk"] Mar 13 07:57:53 crc kubenswrapper[4876]: W0313 07:57:53.746350 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod275d5189_c7c8_44f4_b753_9590b692900c.slice/crio-42a892a031d6418ebe319f0569aa5803c4037f389c131d41f65ddc17a14cb60d WatchSource:0}: Error finding container 42a892a031d6418ebe319f0569aa5803c4037f389c131d41f65ddc17a14cb60d: Status 404 returned error can't find the container with id 42a892a031d6418ebe319f0569aa5803c4037f389c131d41f65ddc17a14cb60d Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.819863 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8359-account-create-update-fr5tj" event={"ID":"32838d97-317e-4102-9fdc-d707329a4f45","Type":"ContainerDied","Data":"9262dd99a449effcbcaa291d8ce9fb5f93a46ad283f7bcae42c6b13718eaab86"} Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.820204 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9262dd99a449effcbcaa291d8ce9fb5f93a46ad283f7bcae42c6b13718eaab86" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.820331 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8359-account-create-update-fr5tj" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.821794 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ccnjk" event={"ID":"275d5189-c7c8-44f4-b753-9590b692900c","Type":"ContainerStarted","Data":"42a892a031d6418ebe319f0569aa5803c4037f389c131d41f65ddc17a14cb60d"} Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.823308 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jkf72" event={"ID":"13197959-749d-4222-ae6c-36fe74c289ff","Type":"ContainerDied","Data":"a49fe40a55dd8de11a7a5ac0b58d5d81368ca46d5e57d56e37265950d3700f0f"} Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.823363 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a49fe40a55dd8de11a7a5ac0b58d5d81368ca46d5e57d56e37265950d3700f0f" Mar 13 07:57:53 crc kubenswrapper[4876]: I0313 07:57:53.823428 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jkf72" Mar 13 07:57:54 crc kubenswrapper[4876]: I0313 07:57:54.995461 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.072912 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-w8nvw"] Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.073534 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" podUID="511f655c-0499-44c9-a654-0094ba185554" containerName="dnsmasq-dns" containerID="cri-o://b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61" gracePeriod=10 Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.356511 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.644114 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-gpnkg"] Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.650783 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-gpnkg"] Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.655681 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.755123 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbp2h\" (UniqueName: \"kubernetes.io/projected/511f655c-0499-44c9-a654-0094ba185554-kube-api-access-cbp2h\") pod \"511f655c-0499-44c9-a654-0094ba185554\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.755261 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-config\") pod \"511f655c-0499-44c9-a654-0094ba185554\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.755328 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-dns-svc\") pod \"511f655c-0499-44c9-a654-0094ba185554\" (UID: \"511f655c-0499-44c9-a654-0094ba185554\") " Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.762481 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/511f655c-0499-44c9-a654-0094ba185554-kube-api-access-cbp2h" (OuterVolumeSpecName: "kube-api-access-cbp2h") pod "511f655c-0499-44c9-a654-0094ba185554" (UID: "511f655c-0499-44c9-a654-0094ba185554"). InnerVolumeSpecName "kube-api-access-cbp2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.795752 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-config" (OuterVolumeSpecName: "config") pod "511f655c-0499-44c9-a654-0094ba185554" (UID: "511f655c-0499-44c9-a654-0094ba185554"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.801991 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "511f655c-0499-44c9-a654-0094ba185554" (UID: "511f655c-0499-44c9-a654-0094ba185554"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.850522 4876 generic.go:334] "Generic (PLEG): container finished" podID="511f655c-0499-44c9-a654-0094ba185554" containerID="b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61" exitCode=0 Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.850579 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" event={"ID":"511f655c-0499-44c9-a654-0094ba185554","Type":"ContainerDied","Data":"b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61"} Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.850611 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" event={"ID":"511f655c-0499-44c9-a654-0094ba185554","Type":"ContainerDied","Data":"a2199d5f27431fda9642b7cbd6afbbe1bae79e526ba8d23cde85f2b99e02ee7b"} Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.850632 4876 scope.go:117] "RemoveContainer" containerID="b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.850759 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-w8nvw" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.859092 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbp2h\" (UniqueName: \"kubernetes.io/projected/511f655c-0499-44c9-a654-0094ba185554-kube-api-access-cbp2h\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.859132 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.859149 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/511f655c-0499-44c9-a654-0094ba185554-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.892570 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-w8nvw"] Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.895397 4876 scope.go:117] "RemoveContainer" containerID="d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.900878 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-w8nvw"] Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.920329 4876 scope.go:117] "RemoveContainer" containerID="b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61" Mar 13 07:57:55 crc kubenswrapper[4876]: E0313 07:57:55.921051 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61\": container with ID starting with b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61 not found: ID does not exist" containerID="b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.921112 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61"} err="failed to get container status \"b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61\": rpc error: code = NotFound desc = could not find container \"b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61\": container with ID starting with b29e61349b02631537a12310d55cdb455ae0015d99565a80a57376c2a3efee61 not found: ID does not exist" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.921148 4876 scope.go:117] "RemoveContainer" containerID="d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d" Mar 13 07:57:55 crc kubenswrapper[4876]: E0313 07:57:55.921688 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d\": container with ID starting with d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d not found: ID does not exist" containerID="d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d" Mar 13 07:57:55 crc kubenswrapper[4876]: I0313 07:57:55.921739 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d"} err="failed to get container status \"d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d\": rpc error: code = NotFound desc = could not find container \"d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d\": container with ID starting with d37ecb1655b280868f619f1da47e4840c5df6a9ad1889c9c3066eb55bc2e7e4d not found: ID does not exist" Mar 13 07:57:56 crc kubenswrapper[4876]: I0313 07:57:56.675477 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:57:56 crc kubenswrapper[4876]: E0313 07:57:56.676951 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 07:57:56 crc kubenswrapper[4876]: E0313 07:57:56.677006 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 07:57:56 crc kubenswrapper[4876]: E0313 07:57:56.677113 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift podName:77efd2bc-d283-4ab6-bd5f-b346b431ea09 nodeName:}" failed. No retries permitted until 2026-03-13 07:58:12.677085378 +0000 UTC m=+1152.347864360 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift") pod "swift-storage-0" (UID: "77efd2bc-d283-4ab6-bd5f-b346b431ea09") : configmap "swift-ring-files" not found Mar 13 07:57:57 crc kubenswrapper[4876]: I0313 07:57:57.050383 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ba6da58-adca-458c-b3e8-004c8be5c1d6" path="/var/lib/kubelet/pods/3ba6da58-adca-458c-b3e8-004c8be5c1d6/volumes" Mar 13 07:57:57 crc kubenswrapper[4876]: I0313 07:57:57.050997 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="511f655c-0499-44c9-a654-0094ba185554" path="/var/lib/kubelet/pods/511f655c-0499-44c9-a654-0094ba185554/volumes" Mar 13 07:57:58 crc kubenswrapper[4876]: I0313 07:57:58.016802 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xrn6k" podUID="3b55481b-c5a5-4602-8f66-c3b002134edb" containerName="ovn-controller" probeResult="failure" output=< Mar 13 07:57:58 crc kubenswrapper[4876]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 07:57:58 crc kubenswrapper[4876]: > Mar 13 07:57:58 crc kubenswrapper[4876]: I0313 07:57:58.880928 4876 generic.go:334] "Generic (PLEG): container finished" podID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerID="8bbb2cd4e0422ba5625aa003e75bef5ff7479286bc579a6d22675536aa504525" exitCode=0 Mar 13 07:57:58 crc kubenswrapper[4876]: I0313 07:57:58.880996 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cca33d7d-7758-4f14-9d61-1785e84ea2b8","Type":"ContainerDied","Data":"8bbb2cd4e0422ba5625aa003e75bef5ff7479286bc579a6d22675536aa504525"} Mar 13 07:57:59 crc kubenswrapper[4876]: I0313 07:57:59.904966 4876 generic.go:334] "Generic (PLEG): container finished" podID="e5294658-a1f0-4340-9d61-c8008a37b5e9" containerID="3b97419c6aa9b29927297319080c64c8ec7e9e8401d46dfc9341853911354a80" exitCode=0 Mar 13 07:57:59 crc kubenswrapper[4876]: I0313 07:57:59.905382 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pbnh2" event={"ID":"e5294658-a1f0-4340-9d61-c8008a37b5e9","Type":"ContainerDied","Data":"3b97419c6aa9b29927297319080c64c8ec7e9e8401d46dfc9341853911354a80"} Mar 13 07:57:59 crc kubenswrapper[4876]: I0313 07:57:59.910780 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cca33d7d-7758-4f14-9d61-1785e84ea2b8","Type":"ContainerStarted","Data":"8d33159caad604faa64957e660287e845db43946fa7402d50ae508453ac633a3"} Mar 13 07:57:59 crc kubenswrapper[4876]: I0313 07:57:59.910973 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:57:59 crc kubenswrapper[4876]: I0313 07:57:59.914697 4876 generic.go:334] "Generic (PLEG): container finished" podID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerID="7bef2f513a2c87a154de6467ad78491d761e7c372e49596ffdb8f85720c39dea" exitCode=0 Mar 13 07:57:59 crc kubenswrapper[4876]: I0313 07:57:59.914736 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678","Type":"ContainerDied","Data":"7bef2f513a2c87a154de6467ad78491d761e7c372e49596ffdb8f85720c39dea"} Mar 13 07:57:59 crc kubenswrapper[4876]: I0313 07:57:59.996497 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=58.116740318 podStartE2EDuration="1m7.99647279s" podCreationTimestamp="2026-03-13 07:56:52 +0000 UTC" firstStartedPulling="2026-03-13 07:57:15.266337063 +0000 UTC m=+1094.937116045" lastFinishedPulling="2026-03-13 07:57:25.146069535 +0000 UTC m=+1104.816848517" observedRunningTime="2026-03-13 07:57:59.987380931 +0000 UTC m=+1139.658159933" watchObservedRunningTime="2026-03-13 07:57:59.99647279 +0000 UTC m=+1139.667251772" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.163145 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556478-vspx7"] Mar 13 07:58:00 crc kubenswrapper[4876]: E0313 07:58:00.164785 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13197959-749d-4222-ae6c-36fe74c289ff" containerName="mariadb-database-create" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.164810 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="13197959-749d-4222-ae6c-36fe74c289ff" containerName="mariadb-database-create" Mar 13 07:58:00 crc kubenswrapper[4876]: E0313 07:58:00.164835 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511f655c-0499-44c9-a654-0094ba185554" containerName="init" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.164846 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="511f655c-0499-44c9-a654-0094ba185554" containerName="init" Mar 13 07:58:00 crc kubenswrapper[4876]: E0313 07:58:00.164871 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32838d97-317e-4102-9fdc-d707329a4f45" containerName="mariadb-account-create-update" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.164882 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="32838d97-317e-4102-9fdc-d707329a4f45" containerName="mariadb-account-create-update" Mar 13 07:58:00 crc kubenswrapper[4876]: E0313 07:58:00.164892 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511f655c-0499-44c9-a654-0094ba185554" containerName="dnsmasq-dns" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.164899 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="511f655c-0499-44c9-a654-0094ba185554" containerName="dnsmasq-dns" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.165080 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="511f655c-0499-44c9-a654-0094ba185554" containerName="dnsmasq-dns" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.165095 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="32838d97-317e-4102-9fdc-d707329a4f45" containerName="mariadb-account-create-update" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.165106 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="13197959-749d-4222-ae6c-36fe74c289ff" containerName="mariadb-database-create" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.165799 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556478-vspx7" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.169371 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.169437 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.171283 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.185268 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556478-vspx7"] Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.344096 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtsbf\" (UniqueName: \"kubernetes.io/projected/99cd306b-1011-48b5-8dfd-b70f6e28f057-kube-api-access-xtsbf\") pod \"auto-csr-approver-29556478-vspx7\" (UID: \"99cd306b-1011-48b5-8dfd-b70f6e28f057\") " pod="openshift-infra/auto-csr-approver-29556478-vspx7" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.446550 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtsbf\" (UniqueName: \"kubernetes.io/projected/99cd306b-1011-48b5-8dfd-b70f6e28f057-kube-api-access-xtsbf\") pod \"auto-csr-approver-29556478-vspx7\" (UID: \"99cd306b-1011-48b5-8dfd-b70f6e28f057\") " pod="openshift-infra/auto-csr-approver-29556478-vspx7" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.467867 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtsbf\" (UniqueName: \"kubernetes.io/projected/99cd306b-1011-48b5-8dfd-b70f6e28f057-kube-api-access-xtsbf\") pod \"auto-csr-approver-29556478-vspx7\" (UID: \"99cd306b-1011-48b5-8dfd-b70f6e28f057\") " pod="openshift-infra/auto-csr-approver-29556478-vspx7" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.496135 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556478-vspx7" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.646314 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-kc84d"] Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.647925 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.650016 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.662309 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-kc84d"] Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.753042 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mls85\" (UniqueName: \"kubernetes.io/projected/c7213beb-842a-408a-8555-57d912b99703-kube-api-access-mls85\") pod \"root-account-create-update-kc84d\" (UID: \"c7213beb-842a-408a-8555-57d912b99703\") " pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.753458 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7213beb-842a-408a-8555-57d912b99703-operator-scripts\") pod \"root-account-create-update-kc84d\" (UID: \"c7213beb-842a-408a-8555-57d912b99703\") " pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.855100 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mls85\" (UniqueName: \"kubernetes.io/projected/c7213beb-842a-408a-8555-57d912b99703-kube-api-access-mls85\") pod \"root-account-create-update-kc84d\" (UID: \"c7213beb-842a-408a-8555-57d912b99703\") " pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.855225 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7213beb-842a-408a-8555-57d912b99703-operator-scripts\") pod \"root-account-create-update-kc84d\" (UID: \"c7213beb-842a-408a-8555-57d912b99703\") " pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.856270 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7213beb-842a-408a-8555-57d912b99703-operator-scripts\") pod \"root-account-create-update-kc84d\" (UID: \"c7213beb-842a-408a-8555-57d912b99703\") " pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.882341 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mls85\" (UniqueName: \"kubernetes.io/projected/c7213beb-842a-408a-8555-57d912b99703-kube-api-access-mls85\") pod \"root-account-create-update-kc84d\" (UID: \"c7213beb-842a-408a-8555-57d912b99703\") " pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:00 crc kubenswrapper[4876]: I0313 07:58:00.971576 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.014260 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xrn6k" podUID="3b55481b-c5a5-4602-8f66-c3b002134edb" containerName="ovn-controller" probeResult="failure" output=< Mar 13 07:58:03 crc kubenswrapper[4876]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 07:58:03 crc kubenswrapper[4876]: > Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.053039 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.062034 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-tb4jh" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.283726 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xrn6k-config-vs66x"] Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.284884 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.302739 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.340762 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xrn6k-config-vs66x"] Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.440096 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-log-ovn\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.440169 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-additional-scripts\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.440190 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t2dm\" (UniqueName: \"kubernetes.io/projected/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-kube-api-access-9t2dm\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.440207 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run-ovn\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.440276 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.440328 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-scripts\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.543019 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-additional-scripts\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.543075 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t2dm\" (UniqueName: \"kubernetes.io/projected/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-kube-api-access-9t2dm\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.543103 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run-ovn\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.543152 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.543202 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-scripts\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.543285 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-log-ovn\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.543656 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-log-ovn\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.543721 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.544201 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run-ovn\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.544768 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-additional-scripts\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.545946 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-scripts\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.568883 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t2dm\" (UniqueName: \"kubernetes.io/projected/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-kube-api-access-9t2dm\") pod \"ovn-controller-xrn6k-config-vs66x\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:03 crc kubenswrapper[4876]: I0313 07:58:03.610889 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.012616 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xrn6k" podUID="3b55481b-c5a5-4602-8f66-c3b002134edb" containerName="ovn-controller" probeResult="failure" output=< Mar 13 07:58:08 crc kubenswrapper[4876]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 07:58:08 crc kubenswrapper[4876]: > Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.399690 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.587004 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-combined-ca-bundle\") pod \"e5294658-a1f0-4340-9d61-c8008a37b5e9\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.587145 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-scripts\") pod \"e5294658-a1f0-4340-9d61-c8008a37b5e9\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.587287 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-swiftconf\") pod \"e5294658-a1f0-4340-9d61-c8008a37b5e9\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.587368 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59ths\" (UniqueName: \"kubernetes.io/projected/e5294658-a1f0-4340-9d61-c8008a37b5e9-kube-api-access-59ths\") pod \"e5294658-a1f0-4340-9d61-c8008a37b5e9\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.587414 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-dispersionconf\") pod \"e5294658-a1f0-4340-9d61-c8008a37b5e9\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.587438 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-ring-data-devices\") pod \"e5294658-a1f0-4340-9d61-c8008a37b5e9\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.587475 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5294658-a1f0-4340-9d61-c8008a37b5e9-etc-swift\") pod \"e5294658-a1f0-4340-9d61-c8008a37b5e9\" (UID: \"e5294658-a1f0-4340-9d61-c8008a37b5e9\") " Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.589374 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5294658-a1f0-4340-9d61-c8008a37b5e9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e5294658-a1f0-4340-9d61-c8008a37b5e9" (UID: "e5294658-a1f0-4340-9d61-c8008a37b5e9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.589748 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e5294658-a1f0-4340-9d61-c8008a37b5e9" (UID: "e5294658-a1f0-4340-9d61-c8008a37b5e9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.605382 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5294658-a1f0-4340-9d61-c8008a37b5e9-kube-api-access-59ths" (OuterVolumeSpecName: "kube-api-access-59ths") pod "e5294658-a1f0-4340-9d61-c8008a37b5e9" (UID: "e5294658-a1f0-4340-9d61-c8008a37b5e9"). InnerVolumeSpecName "kube-api-access-59ths". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.609378 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-scripts" (OuterVolumeSpecName: "scripts") pod "e5294658-a1f0-4340-9d61-c8008a37b5e9" (UID: "e5294658-a1f0-4340-9d61-c8008a37b5e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.611470 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e5294658-a1f0-4340-9d61-c8008a37b5e9" (UID: "e5294658-a1f0-4340-9d61-c8008a37b5e9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.613001 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5294658-a1f0-4340-9d61-c8008a37b5e9" (UID: "e5294658-a1f0-4340-9d61-c8008a37b5e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.634713 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e5294658-a1f0-4340-9d61-c8008a37b5e9" (UID: "e5294658-a1f0-4340-9d61-c8008a37b5e9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.689787 4876 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.689837 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59ths\" (UniqueName: \"kubernetes.io/projected/e5294658-a1f0-4340-9d61-c8008a37b5e9-kube-api-access-59ths\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.689853 4876 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.689866 4876 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.689879 4876 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5294658-a1f0-4340-9d61-c8008a37b5e9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.689890 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5294658-a1f0-4340-9d61-c8008a37b5e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.689902 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5294658-a1f0-4340-9d61-c8008a37b5e9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.816815 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556478-vspx7"] Mar 13 07:58:08 crc kubenswrapper[4876]: W0313 07:58:08.823543 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99cd306b_1011_48b5_8dfd_b70f6e28f057.slice/crio-a86e8a5ef5193feee6876013da2f8cfb9fdff18ddaf05970db3a19a606df9ea8 WatchSource:0}: Error finding container a86e8a5ef5193feee6876013da2f8cfb9fdff18ddaf05970db3a19a606df9ea8: Status 404 returned error can't find the container with id a86e8a5ef5193feee6876013da2f8cfb9fdff18ddaf05970db3a19a606df9ea8 Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.915140 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xrn6k-config-vs66x"] Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.934317 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-kc84d"] Mar 13 07:58:08 crc kubenswrapper[4876]: W0313 07:58:08.942022 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7213beb_842a_408a_8555_57d912b99703.slice/crio-f488d4b2888f0c7c2ddfa19eab88feaed2d2c0cf96c97c28c5e81c72ef39bb59 WatchSource:0}: Error finding container f488d4b2888f0c7c2ddfa19eab88feaed2d2c0cf96c97c28c5e81c72ef39bb59: Status 404 returned error can't find the container with id f488d4b2888f0c7c2ddfa19eab88feaed2d2c0cf96c97c28c5e81c72ef39bb59 Mar 13 07:58:08 crc kubenswrapper[4876]: I0313 07:58:08.950588 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.003410 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xrn6k-config-vs66x" event={"ID":"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96","Type":"ContainerStarted","Data":"42bce05f09295cf05b3919d0aaf80e3bc8db465358822bb454d12baf8bcf9fd6"} Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.009585 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678","Type":"ContainerStarted","Data":"7a2e4794cd7b4ce4ab5ad2b4921db1d2e8ca47414faee7c132f8ec81ad3a2c3c"} Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.009880 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.010834 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556478-vspx7" event={"ID":"99cd306b-1011-48b5-8dfd-b70f6e28f057","Type":"ContainerStarted","Data":"a86e8a5ef5193feee6876013da2f8cfb9fdff18ddaf05970db3a19a606df9ea8"} Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.012808 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-kc84d" event={"ID":"c7213beb-842a-408a-8555-57d912b99703","Type":"ContainerStarted","Data":"f488d4b2888f0c7c2ddfa19eab88feaed2d2c0cf96c97c28c5e81c72ef39bb59"} Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.015155 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pbnh2" event={"ID":"e5294658-a1f0-4340-9d61-c8008a37b5e9","Type":"ContainerDied","Data":"104b1d585cfafe6c568d3d9bf9d008b9540d8fc06483dd249832520dbeff227d"} Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.015184 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="104b1d585cfafe6c568d3d9bf9d008b9540d8fc06483dd249832520dbeff227d" Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.015315 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pbnh2" Mar 13 07:58:09 crc kubenswrapper[4876]: I0313 07:58:09.037634 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=65.799110333 podStartE2EDuration="1m16.037612575s" podCreationTimestamp="2026-03-13 07:56:53 +0000 UTC" firstStartedPulling="2026-03-13 07:57:14.97008476 +0000 UTC m=+1094.640863742" lastFinishedPulling="2026-03-13 07:57:25.208587002 +0000 UTC m=+1104.879365984" observedRunningTime="2026-03-13 07:58:09.035801204 +0000 UTC m=+1148.706580186" watchObservedRunningTime="2026-03-13 07:58:09.037612575 +0000 UTC m=+1148.708391557" Mar 13 07:58:10 crc kubenswrapper[4876]: I0313 07:58:10.027344 4876 generic.go:334] "Generic (PLEG): container finished" podID="a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" containerID="0054e18eed70fb08e7510798b681c205692d43652c7d8a318ba96752774ce19c" exitCode=0 Mar 13 07:58:10 crc kubenswrapper[4876]: I0313 07:58:10.027435 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xrn6k-config-vs66x" event={"ID":"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96","Type":"ContainerDied","Data":"0054e18eed70fb08e7510798b681c205692d43652c7d8a318ba96752774ce19c"} Mar 13 07:58:10 crc kubenswrapper[4876]: I0313 07:58:10.033452 4876 generic.go:334] "Generic (PLEG): container finished" podID="c7213beb-842a-408a-8555-57d912b99703" containerID="4f9013fb4dc8dee9f4e68587e3f4df7c9ea1d9d5752bd42e435e22f98ff21602" exitCode=0 Mar 13 07:58:10 crc kubenswrapper[4876]: I0313 07:58:10.033569 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-kc84d" event={"ID":"c7213beb-842a-408a-8555-57d912b99703","Type":"ContainerDied","Data":"4f9013fb4dc8dee9f4e68587e3f4df7c9ea1d9d5752bd42e435e22f98ff21602"} Mar 13 07:58:10 crc kubenswrapper[4876]: I0313 07:58:10.036167 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ccnjk" event={"ID":"275d5189-c7c8-44f4-b753-9590b692900c","Type":"ContainerStarted","Data":"5d3759610c114fd90a7e5012fd273548fce3da5c843f4daf17c879cb01455d7d"} Mar 13 07:58:10 crc kubenswrapper[4876]: I0313 07:58:10.108388 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-ccnjk" podStartSLOduration=3.472855879 podStartE2EDuration="18.1083402s" podCreationTimestamp="2026-03-13 07:57:52 +0000 UTC" firstStartedPulling="2026-03-13 07:57:53.749940842 +0000 UTC m=+1133.420719834" lastFinishedPulling="2026-03-13 07:58:08.385425183 +0000 UTC m=+1148.056204155" observedRunningTime="2026-03-13 07:58:10.098276405 +0000 UTC m=+1149.769055387" watchObservedRunningTime="2026-03-13 07:58:10.1083402 +0000 UTC m=+1149.779119212" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.075445 4876 generic.go:334] "Generic (PLEG): container finished" podID="99cd306b-1011-48b5-8dfd-b70f6e28f057" containerID="d9c45b3607ae9397dd7932a99ff50e2d3ef797c6a4f714c8947b7859a18a262e" exitCode=0 Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.075586 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556478-vspx7" event={"ID":"99cd306b-1011-48b5-8dfd-b70f6e28f057","Type":"ContainerDied","Data":"d9c45b3607ae9397dd7932a99ff50e2d3ef797c6a4f714c8947b7859a18a262e"} Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.425775 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.522517 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.569873 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-scripts\") pod \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570065 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-log-ovn\") pod \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570100 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-additional-scripts\") pod \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570195 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run\") pod \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570225 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" (UID: "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570323 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run" (OuterVolumeSpecName: "var-run") pod "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" (UID: "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570500 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t2dm\" (UniqueName: \"kubernetes.io/projected/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-kube-api-access-9t2dm\") pod \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570546 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run-ovn\") pod \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\" (UID: \"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96\") " Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570697 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" (UID: "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.570999 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" (UID: "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.571119 4876 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.571145 4876 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.571158 4876 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.571600 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-scripts" (OuterVolumeSpecName: "scripts") pod "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" (UID: "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.579420 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-kube-api-access-9t2dm" (OuterVolumeSpecName: "kube-api-access-9t2dm") pod "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" (UID: "a7ba2a22-13ba-4dfd-86ef-ac4e9097da96"). InnerVolumeSpecName "kube-api-access-9t2dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.672385 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7213beb-842a-408a-8555-57d912b99703-operator-scripts\") pod \"c7213beb-842a-408a-8555-57d912b99703\" (UID: \"c7213beb-842a-408a-8555-57d912b99703\") " Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.672977 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mls85\" (UniqueName: \"kubernetes.io/projected/c7213beb-842a-408a-8555-57d912b99703-kube-api-access-mls85\") pod \"c7213beb-842a-408a-8555-57d912b99703\" (UID: \"c7213beb-842a-408a-8555-57d912b99703\") " Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.673585 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7213beb-842a-408a-8555-57d912b99703-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7213beb-842a-408a-8555-57d912b99703" (UID: "c7213beb-842a-408a-8555-57d912b99703"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.674998 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.675147 4876 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.675215 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7213beb-842a-408a-8555-57d912b99703-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.675290 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t2dm\" (UniqueName: \"kubernetes.io/projected/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96-kube-api-access-9t2dm\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.677598 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7213beb-842a-408a-8555-57d912b99703-kube-api-access-mls85" (OuterVolumeSpecName: "kube-api-access-mls85") pod "c7213beb-842a-408a-8555-57d912b99703" (UID: "c7213beb-842a-408a-8555-57d912b99703"). InnerVolumeSpecName "kube-api-access-mls85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:11 crc kubenswrapper[4876]: I0313 07:58:11.777159 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mls85\" (UniqueName: \"kubernetes.io/projected/c7213beb-842a-408a-8555-57d912b99703-kube-api-access-mls85\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.086838 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xrn6k-config-vs66x" event={"ID":"a7ba2a22-13ba-4dfd-86ef-ac4e9097da96","Type":"ContainerDied","Data":"42bce05f09295cf05b3919d0aaf80e3bc8db465358822bb454d12baf8bcf9fd6"} Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.086953 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42bce05f09295cf05b3919d0aaf80e3bc8db465358822bb454d12baf8bcf9fd6" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.086870 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xrn6k-config-vs66x" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.091405 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-kc84d" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.092436 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-kc84d" event={"ID":"c7213beb-842a-408a-8555-57d912b99703","Type":"ContainerDied","Data":"f488d4b2888f0c7c2ddfa19eab88feaed2d2c0cf96c97c28c5e81c72ef39bb59"} Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.092495 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f488d4b2888f0c7c2ddfa19eab88feaed2d2c0cf96c97c28c5e81c72ef39bb59" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.430787 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556478-vspx7" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.554876 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xrn6k-config-vs66x"] Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.564572 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xrn6k-config-vs66x"] Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.592086 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtsbf\" (UniqueName: \"kubernetes.io/projected/99cd306b-1011-48b5-8dfd-b70f6e28f057-kube-api-access-xtsbf\") pod \"99cd306b-1011-48b5-8dfd-b70f6e28f057\" (UID: \"99cd306b-1011-48b5-8dfd-b70f6e28f057\") " Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.598435 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99cd306b-1011-48b5-8dfd-b70f6e28f057-kube-api-access-xtsbf" (OuterVolumeSpecName: "kube-api-access-xtsbf") pod "99cd306b-1011-48b5-8dfd-b70f6e28f057" (UID: "99cd306b-1011-48b5-8dfd-b70f6e28f057"). InnerVolumeSpecName "kube-api-access-xtsbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.694670 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.695116 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtsbf\" (UniqueName: \"kubernetes.io/projected/99cd306b-1011-48b5-8dfd-b70f6e28f057-kube-api-access-xtsbf\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.702315 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77efd2bc-d283-4ab6-bd5f-b346b431ea09-etc-swift\") pod \"swift-storage-0\" (UID: \"77efd2bc-d283-4ab6-bd5f-b346b431ea09\") " pod="openstack/swift-storage-0" Mar 13 07:58:12 crc kubenswrapper[4876]: I0313 07:58:12.921276 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.027629 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-xrn6k" Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.088732 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" path="/var/lib/kubelet/pods/a7ba2a22-13ba-4dfd-86ef-ac4e9097da96/volumes" Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.102671 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556478-vspx7" event={"ID":"99cd306b-1011-48b5-8dfd-b70f6e28f057","Type":"ContainerDied","Data":"a86e8a5ef5193feee6876013da2f8cfb9fdff18ddaf05970db3a19a606df9ea8"} Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.102718 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a86e8a5ef5193feee6876013da2f8cfb9fdff18ddaf05970db3a19a606df9ea8" Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.102781 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556478-vspx7" Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.498304 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556472-68q8k"] Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.506144 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556472-68q8k"] Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.600579 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 07:58:13 crc kubenswrapper[4876]: I0313 07:58:13.607349 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 07:58:14 crc kubenswrapper[4876]: I0313 07:58:14.115849 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"fda09f57a182a1319c63fd4f7c4b7a1992532453e858da475a261d955473cffc"} Mar 13 07:58:14 crc kubenswrapper[4876]: I0313 07:58:14.167842 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 07:58:15 crc kubenswrapper[4876]: I0313 07:58:15.046452 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="710597ba-1b67-45a8-83de-8c9a23cccd4c" path="/var/lib/kubelet/pods/710597ba-1b67-45a8-83de-8c9a23cccd4c/volumes" Mar 13 07:58:16 crc kubenswrapper[4876]: I0313 07:58:16.136427 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"208b5df2bf970c83ce2b5574a1f5d16ee4605421a351c7aa80edf14d4c752a0d"} Mar 13 07:58:16 crc kubenswrapper[4876]: I0313 07:58:16.137077 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"4d5dd65a507b01a5c2dacdabff98b59d51128aaab4cbfae94d6ae2a5193014f5"} Mar 13 07:58:16 crc kubenswrapper[4876]: I0313 07:58:16.137090 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"6e439da17302dee26b28e82b627b1530ddc87f5f6b03f8e20b042818bba358ca"} Mar 13 07:58:16 crc kubenswrapper[4876]: I0313 07:58:16.137100 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"16c6b7b0b51b055c8d6b5d6b0655a65a7cb8157948a085305c1cebbf96c7afd1"} Mar 13 07:58:18 crc kubenswrapper[4876]: I0313 07:58:18.156053 4876 generic.go:334] "Generic (PLEG): container finished" podID="275d5189-c7c8-44f4-b753-9590b692900c" containerID="5d3759610c114fd90a7e5012fd273548fce3da5c843f4daf17c879cb01455d7d" exitCode=0 Mar 13 07:58:18 crc kubenswrapper[4876]: I0313 07:58:18.156129 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ccnjk" event={"ID":"275d5189-c7c8-44f4-b753-9590b692900c","Type":"ContainerDied","Data":"5d3759610c114fd90a7e5012fd273548fce3da5c843f4daf17c879cb01455d7d"} Mar 13 07:58:18 crc kubenswrapper[4876]: I0313 07:58:18.163318 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"ad7093b38e0f1becb8133a496c5061ce509f348dd529bf34adf4f23dae5625ee"} Mar 13 07:58:18 crc kubenswrapper[4876]: I0313 07:58:18.163386 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"b06ab0a0c07f6f8e6b3c218a92d5663f6d12bbc1c0f53e532d71aa54c070304e"} Mar 13 07:58:18 crc kubenswrapper[4876]: I0313 07:58:18.163401 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"f970b546c2efa05eedbf8dfc1f45dd37e5570f508fafcf23c3975788b1e331c8"} Mar 13 07:58:18 crc kubenswrapper[4876]: I0313 07:58:18.163414 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"83554d8939bd5d86ad689e2e5ed48919f967cb7a1aed34d4099b4369324719ab"} Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.785482 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ccnjk" Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.928398 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-combined-ca-bundle\") pod \"275d5189-c7c8-44f4-b753-9590b692900c\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.928541 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-db-sync-config-data\") pod \"275d5189-c7c8-44f4-b753-9590b692900c\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.928582 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-config-data\") pod \"275d5189-c7c8-44f4-b753-9590b692900c\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.928725 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhvvl\" (UniqueName: \"kubernetes.io/projected/275d5189-c7c8-44f4-b753-9590b692900c-kube-api-access-bhvvl\") pod \"275d5189-c7c8-44f4-b753-9590b692900c\" (UID: \"275d5189-c7c8-44f4-b753-9590b692900c\") " Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.934874 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/275d5189-c7c8-44f4-b753-9590b692900c-kube-api-access-bhvvl" (OuterVolumeSpecName: "kube-api-access-bhvvl") pod "275d5189-c7c8-44f4-b753-9590b692900c" (UID: "275d5189-c7c8-44f4-b753-9590b692900c"). InnerVolumeSpecName "kube-api-access-bhvvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.939337 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "275d5189-c7c8-44f4-b753-9590b692900c" (UID: "275d5189-c7c8-44f4-b753-9590b692900c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.961399 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "275d5189-c7c8-44f4-b753-9590b692900c" (UID: "275d5189-c7c8-44f4-b753-9590b692900c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:19 crc kubenswrapper[4876]: I0313 07:58:19.973104 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-config-data" (OuterVolumeSpecName: "config-data") pod "275d5189-c7c8-44f4-b753-9590b692900c" (UID: "275d5189-c7c8-44f4-b753-9590b692900c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.030600 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhvvl\" (UniqueName: \"kubernetes.io/projected/275d5189-c7c8-44f4-b753-9590b692900c-kube-api-access-bhvvl\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.030641 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.030650 4876 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.030659 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/275d5189-c7c8-44f4-b753-9590b692900c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.187120 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"fc3881a4778b7c276778668ae0a016389307e8ac0af19d8fcae66f7188e6afa7"} Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.187182 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"d932bc1643002d1316c798f6f251e18dff5fed85937730cdd6e6c37186aed31d"} Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.187198 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"0ef436ecd61c4f6e0602ce820b1f94802735fe68929535a1ad80c2ab1ca7c6c0"} Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.187210 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"6acfbb635e056cbaf38e25bc03f9b2093a51e4e1fd883c30c81f40c3744f6978"} Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.189367 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ccnjk" event={"ID":"275d5189-c7c8-44f4-b753-9590b692900c","Type":"ContainerDied","Data":"42a892a031d6418ebe319f0569aa5803c4037f389c131d41f65ddc17a14cb60d"} Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.189423 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42a892a031d6418ebe319f0569aa5803c4037f389c131d41f65ddc17a14cb60d" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.189653 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ccnjk" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.704010 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jfbrh"] Mar 13 07:58:20 crc kubenswrapper[4876]: E0313 07:58:20.705292 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7213beb-842a-408a-8555-57d912b99703" containerName="mariadb-account-create-update" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705309 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7213beb-842a-408a-8555-57d912b99703" containerName="mariadb-account-create-update" Mar 13 07:58:20 crc kubenswrapper[4876]: E0313 07:58:20.705360 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5294658-a1f0-4340-9d61-c8008a37b5e9" containerName="swift-ring-rebalance" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705368 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5294658-a1f0-4340-9d61-c8008a37b5e9" containerName="swift-ring-rebalance" Mar 13 07:58:20 crc kubenswrapper[4876]: E0313 07:58:20.705406 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" containerName="ovn-config" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705414 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" containerName="ovn-config" Mar 13 07:58:20 crc kubenswrapper[4876]: E0313 07:58:20.705429 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275d5189-c7c8-44f4-b753-9590b692900c" containerName="glance-db-sync" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705435 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="275d5189-c7c8-44f4-b753-9590b692900c" containerName="glance-db-sync" Mar 13 07:58:20 crc kubenswrapper[4876]: E0313 07:58:20.705447 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99cd306b-1011-48b5-8dfd-b70f6e28f057" containerName="oc" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705454 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="99cd306b-1011-48b5-8dfd-b70f6e28f057" containerName="oc" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705762 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5294658-a1f0-4340-9d61-c8008a37b5e9" containerName="swift-ring-rebalance" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705781 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7213beb-842a-408a-8555-57d912b99703" containerName="mariadb-account-create-update" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705797 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="99cd306b-1011-48b5-8dfd-b70f6e28f057" containerName="oc" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705810 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="275d5189-c7c8-44f4-b753-9590b692900c" containerName="glance-db-sync" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.705827 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ba2a22-13ba-4dfd-86ef-ac4e9097da96" containerName="ovn-config" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.707749 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.748027 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jfbrh"] Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.851294 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tngvt\" (UniqueName: \"kubernetes.io/projected/c480a919-d0a1-47e1-96cf-a37fbd8d3962-kube-api-access-tngvt\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.851394 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.851458 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.851541 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.851624 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-config\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.953362 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.953450 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.953504 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-config\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.953529 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tngvt\" (UniqueName: \"kubernetes.io/projected/c480a919-d0a1-47e1-96cf-a37fbd8d3962-kube-api-access-tngvt\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.953567 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.954531 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.955363 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.955405 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-config\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.955771 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:20 crc kubenswrapper[4876]: I0313 07:58:20.981884 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tngvt\" (UniqueName: \"kubernetes.io/projected/c480a919-d0a1-47e1-96cf-a37fbd8d3962-kube-api-access-tngvt\") pod \"dnsmasq-dns-5b946c75cc-jfbrh\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.089737 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.221189 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"f2390fc24b77c15a93c6d3c79c32d378bc63a9828e7dbe597d23ea26607eb6f8"} Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.221742 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"3d10ba6e2c6af4e2c208ec31b78ffc40fd2a3eceab65b4a658483026ee1742b9"} Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.221760 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77efd2bc-d283-4ab6-bd5f-b346b431ea09","Type":"ContainerStarted","Data":"ae92cadb0e5fe90a26f6ca02294185acf9d02399dfaaab541a8e0eae5fc23022"} Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.264954 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.6218865 podStartE2EDuration="42.264925958s" podCreationTimestamp="2026-03-13 07:57:39 +0000 UTC" firstStartedPulling="2026-03-13 07:58:13.600387868 +0000 UTC m=+1153.271166850" lastFinishedPulling="2026-03-13 07:58:19.243427326 +0000 UTC m=+1158.914206308" observedRunningTime="2026-03-13 07:58:21.260974895 +0000 UTC m=+1160.931753877" watchObservedRunningTime="2026-03-13 07:58:21.264925958 +0000 UTC m=+1160.935704940" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.576191 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jfbrh"] Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.619639 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-crr5d"] Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.621749 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.636591 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.648398 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-crr5d"] Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.693873 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jfbrh"] Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.773545 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.774191 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7zpq\" (UniqueName: \"kubernetes.io/projected/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-kube-api-access-z7zpq\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.774283 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-config\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.774329 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.774367 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.774536 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.877169 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.877261 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.877345 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.877372 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7zpq\" (UniqueName: \"kubernetes.io/projected/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-kube-api-access-z7zpq\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.877457 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-config\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.877503 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.877550 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.878222 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.879280 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.879564 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-config\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.879605 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.900209 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7zpq\" (UniqueName: \"kubernetes.io/projected/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-kube-api-access-z7zpq\") pod \"dnsmasq-dns-74f6bcbc87-crr5d\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:21 crc kubenswrapper[4876]: I0313 07:58:21.949537 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.248074 4876 generic.go:334] "Generic (PLEG): container finished" podID="c480a919-d0a1-47e1-96cf-a37fbd8d3962" containerID="6162ad2aaed8457a460111d57be8472fb99d261d89e21b78649f72109fbf80f3" exitCode=0 Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.249420 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" event={"ID":"c480a919-d0a1-47e1-96cf-a37fbd8d3962","Type":"ContainerDied","Data":"6162ad2aaed8457a460111d57be8472fb99d261d89e21b78649f72109fbf80f3"} Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.249460 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" event={"ID":"c480a919-d0a1-47e1-96cf-a37fbd8d3962","Type":"ContainerStarted","Data":"1c7e7c2bfaa0b7cb0f668b051759d0fdf4a5ca2fb8fdc81fa03188c71c03ad78"} Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.477767 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-crr5d"] Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.668293 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.803103 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-dns-svc\") pod \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.803284 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-sb\") pod \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.803349 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-config\") pod \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.803453 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-nb\") pod \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.803491 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tngvt\" (UniqueName: \"kubernetes.io/projected/c480a919-d0a1-47e1-96cf-a37fbd8d3962-kube-api-access-tngvt\") pod \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\" (UID: \"c480a919-d0a1-47e1-96cf-a37fbd8d3962\") " Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.811593 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c480a919-d0a1-47e1-96cf-a37fbd8d3962-kube-api-access-tngvt" (OuterVolumeSpecName: "kube-api-access-tngvt") pod "c480a919-d0a1-47e1-96cf-a37fbd8d3962" (UID: "c480a919-d0a1-47e1-96cf-a37fbd8d3962"). InnerVolumeSpecName "kube-api-access-tngvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.827255 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-config" (OuterVolumeSpecName: "config") pod "c480a919-d0a1-47e1-96cf-a37fbd8d3962" (UID: "c480a919-d0a1-47e1-96cf-a37fbd8d3962"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.827563 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c480a919-d0a1-47e1-96cf-a37fbd8d3962" (UID: "c480a919-d0a1-47e1-96cf-a37fbd8d3962"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.828456 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c480a919-d0a1-47e1-96cf-a37fbd8d3962" (UID: "c480a919-d0a1-47e1-96cf-a37fbd8d3962"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.830880 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c480a919-d0a1-47e1-96cf-a37fbd8d3962" (UID: "c480a919-d0a1-47e1-96cf-a37fbd8d3962"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.905937 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.906003 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.906019 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.906036 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tngvt\" (UniqueName: \"kubernetes.io/projected/c480a919-d0a1-47e1-96cf-a37fbd8d3962-kube-api-access-tngvt\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:22 crc kubenswrapper[4876]: I0313 07:58:22.906050 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c480a919-d0a1-47e1-96cf-a37fbd8d3962-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:23 crc kubenswrapper[4876]: I0313 07:58:23.260435 4876 generic.go:334] "Generic (PLEG): container finished" podID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerID="45fc2002c0203cc2e54785692fc0ba68400f3f50db1ce89d4c4fade21ed6ac0e" exitCode=0 Mar 13 07:58:23 crc kubenswrapper[4876]: I0313 07:58:23.260843 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" event={"ID":"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6","Type":"ContainerDied","Data":"45fc2002c0203cc2e54785692fc0ba68400f3f50db1ce89d4c4fade21ed6ac0e"} Mar 13 07:58:23 crc kubenswrapper[4876]: I0313 07:58:23.260878 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" event={"ID":"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6","Type":"ContainerStarted","Data":"f072db7b78651a7b2d7a09b36e14ce6bc1ab7abe9f4259888ba0a2a95036b139"} Mar 13 07:58:23 crc kubenswrapper[4876]: I0313 07:58:23.273843 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" event={"ID":"c480a919-d0a1-47e1-96cf-a37fbd8d3962","Type":"ContainerDied","Data":"1c7e7c2bfaa0b7cb0f668b051759d0fdf4a5ca2fb8fdc81fa03188c71c03ad78"} Mar 13 07:58:23 crc kubenswrapper[4876]: I0313 07:58:23.273932 4876 scope.go:117] "RemoveContainer" containerID="6162ad2aaed8457a460111d57be8472fb99d261d89e21b78649f72109fbf80f3" Mar 13 07:58:23 crc kubenswrapper[4876]: I0313 07:58:23.274082 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jfbrh" Mar 13 07:58:23 crc kubenswrapper[4876]: I0313 07:58:23.354685 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jfbrh"] Mar 13 07:58:23 crc kubenswrapper[4876]: I0313 07:58:23.361897 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jfbrh"] Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.285586 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" event={"ID":"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6","Type":"ContainerStarted","Data":"4fd6fad89a28c78e7cd9a7afe90312ce4fbec880252f8ff385bc6a6bf26aefe0"} Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.286158 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.312154 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" podStartSLOduration=3.312124614 podStartE2EDuration="3.312124614s" podCreationTimestamp="2026-03-13 07:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:24.307905604 +0000 UTC m=+1163.978684626" watchObservedRunningTime="2026-03-13 07:58:24.312124614 +0000 UTC m=+1163.982903596" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.537473 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.607681 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.607754 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.825380 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-67gcf"] Mar 13 07:58:24 crc kubenswrapper[4876]: E0313 07:58:24.825798 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c480a919-d0a1-47e1-96cf-a37fbd8d3962" containerName="init" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.825819 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c480a919-d0a1-47e1-96cf-a37fbd8d3962" containerName="init" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.825998 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c480a919-d0a1-47e1-96cf-a37fbd8d3962" containerName="init" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.826596 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.838643 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-67gcf"] Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.959276 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5253-account-create-update-zqf6n"] Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.960438 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.964501 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.970206 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5253-account-create-update-zqf6n"] Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.996643 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b06a84e-50b9-4e3f-b57e-4db1b023699d-operator-scripts\") pod \"cinder-db-create-67gcf\" (UID: \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\") " pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:24 crc kubenswrapper[4876]: I0313 07:58:24.996821 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swbrj\" (UniqueName: \"kubernetes.io/projected/0b06a84e-50b9-4e3f-b57e-4db1b023699d-kube-api-access-swbrj\") pod \"cinder-db-create-67gcf\" (UID: \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\") " pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.047812 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c480a919-d0a1-47e1-96cf-a37fbd8d3962" path="/var/lib/kubelet/pods/c480a919-d0a1-47e1-96cf-a37fbd8d3962/volumes" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.098545 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b06a84e-50b9-4e3f-b57e-4db1b023699d-operator-scripts\") pod \"cinder-db-create-67gcf\" (UID: \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\") " pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.098657 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e16d2f43-d740-4edb-b41d-806e8bb28ff0-operator-scripts\") pod \"cinder-5253-account-create-update-zqf6n\" (UID: \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\") " pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.098738 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swbrj\" (UniqueName: \"kubernetes.io/projected/0b06a84e-50b9-4e3f-b57e-4db1b023699d-kube-api-access-swbrj\") pod \"cinder-db-create-67gcf\" (UID: \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\") " pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.098934 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwczk\" (UniqueName: \"kubernetes.io/projected/e16d2f43-d740-4edb-b41d-806e8bb28ff0-kube-api-access-qwczk\") pod \"cinder-5253-account-create-update-zqf6n\" (UID: \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\") " pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.099612 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b06a84e-50b9-4e3f-b57e-4db1b023699d-operator-scripts\") pod \"cinder-db-create-67gcf\" (UID: \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\") " pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.134218 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-xwt9g"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.136366 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swbrj\" (UniqueName: \"kubernetes.io/projected/0b06a84e-50b9-4e3f-b57e-4db1b023699d-kube-api-access-swbrj\") pod \"cinder-db-create-67gcf\" (UID: \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\") " pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.149024 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.157863 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xwt9g"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.158010 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.203262 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwczk\" (UniqueName: \"kubernetes.io/projected/e16d2f43-d740-4edb-b41d-806e8bb28ff0-kube-api-access-qwczk\") pod \"cinder-5253-account-create-update-zqf6n\" (UID: \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\") " pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.203752 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e16d2f43-d740-4edb-b41d-806e8bb28ff0-operator-scripts\") pod \"cinder-5253-account-create-update-zqf6n\" (UID: \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\") " pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.205228 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e16d2f43-d740-4edb-b41d-806e8bb28ff0-operator-scripts\") pod \"cinder-5253-account-create-update-zqf6n\" (UID: \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\") " pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.227033 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwczk\" (UniqueName: \"kubernetes.io/projected/e16d2f43-d740-4edb-b41d-806e8bb28ff0-kube-api-access-qwczk\") pod \"cinder-5253-account-create-update-zqf6n\" (UID: \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\") " pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.229504 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-jm9nf"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.231126 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.265865 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jm9nf"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.278965 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.311783 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-operator-scripts\") pod \"barbican-db-create-xwt9g\" (UID: \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\") " pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.312025 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwbfl\" (UniqueName: \"kubernetes.io/projected/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-kube-api-access-gwbfl\") pod \"barbican-db-create-xwt9g\" (UID: \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\") " pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.394469 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-45rb2"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.395996 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.404032 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-btt8h" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.404292 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.404467 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.404582 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.415947 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-operator-scripts\") pod \"barbican-db-create-xwt9g\" (UID: \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\") " pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.416114 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwbfl\" (UniqueName: \"kubernetes.io/projected/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-kube-api-access-gwbfl\") pod \"barbican-db-create-xwt9g\" (UID: \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\") " pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.416146 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f174ba-30f6-4440-9d1c-037d413b5f90-operator-scripts\") pod \"neutron-db-create-jm9nf\" (UID: \"44f174ba-30f6-4440-9d1c-037d413b5f90\") " pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.416214 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf7kz\" (UniqueName: \"kubernetes.io/projected/44f174ba-30f6-4440-9d1c-037d413b5f90-kube-api-access-mf7kz\") pod \"neutron-db-create-jm9nf\" (UID: \"44f174ba-30f6-4440-9d1c-037d413b5f90\") " pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.420142 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-operator-scripts\") pod \"barbican-db-create-xwt9g\" (UID: \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\") " pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.446436 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-45rb2"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.461158 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwbfl\" (UniqueName: \"kubernetes.io/projected/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-kube-api-access-gwbfl\") pod \"barbican-db-create-xwt9g\" (UID: \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\") " pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.483302 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a6be-account-create-update-z9kh5"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.485012 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.487530 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.494508 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a6be-account-create-update-z9kh5"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.520670 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf7kz\" (UniqueName: \"kubernetes.io/projected/44f174ba-30f6-4440-9d1c-037d413b5f90-kube-api-access-mf7kz\") pod \"neutron-db-create-jm9nf\" (UID: \"44f174ba-30f6-4440-9d1c-037d413b5f90\") " pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.520789 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-combined-ca-bundle\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.520877 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbh6k\" (UniqueName: \"kubernetes.io/projected/8523dd40-992c-4540-8bb0-2464da42a16c-kube-api-access-qbh6k\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.520932 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-config-data\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.520966 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f174ba-30f6-4440-9d1c-037d413b5f90-operator-scripts\") pod \"neutron-db-create-jm9nf\" (UID: \"44f174ba-30f6-4440-9d1c-037d413b5f90\") " pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.521560 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c123-account-create-update-r29rm"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.522492 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f174ba-30f6-4440-9d1c-037d413b5f90-operator-scripts\") pod \"neutron-db-create-jm9nf\" (UID: \"44f174ba-30f6-4440-9d1c-037d413b5f90\") " pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.526789 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.529583 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c123-account-create-update-r29rm"] Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.531141 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.547097 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf7kz\" (UniqueName: \"kubernetes.io/projected/44f174ba-30f6-4440-9d1c-037d413b5f90-kube-api-access-mf7kz\") pod \"neutron-db-create-jm9nf\" (UID: \"44f174ba-30f6-4440-9d1c-037d413b5f90\") " pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.622521 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbh6k\" (UniqueName: \"kubernetes.io/projected/8523dd40-992c-4540-8bb0-2464da42a16c-kube-api-access-qbh6k\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.622615 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e28a17-18b9-44a3-beec-cf688f768519-operator-scripts\") pod \"barbican-a6be-account-create-update-z9kh5\" (UID: \"52e28a17-18b9-44a3-beec-cf688f768519\") " pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.622660 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-config-data\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.622698 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p977\" (UniqueName: \"kubernetes.io/projected/27f72293-3842-46ab-87ad-fe2345ea4b3a-kube-api-access-7p977\") pod \"neutron-c123-account-create-update-r29rm\" (UID: \"27f72293-3842-46ab-87ad-fe2345ea4b3a\") " pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.622759 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27f72293-3842-46ab-87ad-fe2345ea4b3a-operator-scripts\") pod \"neutron-c123-account-create-update-r29rm\" (UID: \"27f72293-3842-46ab-87ad-fe2345ea4b3a\") " pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.622832 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-combined-ca-bundle\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.622877 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmmrh\" (UniqueName: \"kubernetes.io/projected/52e28a17-18b9-44a3-beec-cf688f768519-kube-api-access-gmmrh\") pod \"barbican-a6be-account-create-update-z9kh5\" (UID: \"52e28a17-18b9-44a3-beec-cf688f768519\") " pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.626438 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-combined-ca-bundle\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.626906 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-config-data\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.640131 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbh6k\" (UniqueName: \"kubernetes.io/projected/8523dd40-992c-4540-8bb0-2464da42a16c-kube-api-access-qbh6k\") pod \"keystone-db-sync-45rb2\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:25 crc kubenswrapper[4876]: I0313 07:58:25.666985 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.702886 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.724785 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmmrh\" (UniqueName: \"kubernetes.io/projected/52e28a17-18b9-44a3-beec-cf688f768519-kube-api-access-gmmrh\") pod \"barbican-a6be-account-create-update-z9kh5\" (UID: \"52e28a17-18b9-44a3-beec-cf688f768519\") " pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.724856 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e28a17-18b9-44a3-beec-cf688f768519-operator-scripts\") pod \"barbican-a6be-account-create-update-z9kh5\" (UID: \"52e28a17-18b9-44a3-beec-cf688f768519\") " pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.724898 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p977\" (UniqueName: \"kubernetes.io/projected/27f72293-3842-46ab-87ad-fe2345ea4b3a-kube-api-access-7p977\") pod \"neutron-c123-account-create-update-r29rm\" (UID: \"27f72293-3842-46ab-87ad-fe2345ea4b3a\") " pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.724924 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27f72293-3842-46ab-87ad-fe2345ea4b3a-operator-scripts\") pod \"neutron-c123-account-create-update-r29rm\" (UID: \"27f72293-3842-46ab-87ad-fe2345ea4b3a\") " pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.725915 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27f72293-3842-46ab-87ad-fe2345ea4b3a-operator-scripts\") pod \"neutron-c123-account-create-update-r29rm\" (UID: \"27f72293-3842-46ab-87ad-fe2345ea4b3a\") " pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.726885 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e28a17-18b9-44a3-beec-cf688f768519-operator-scripts\") pod \"barbican-a6be-account-create-update-z9kh5\" (UID: \"52e28a17-18b9-44a3-beec-cf688f768519\") " pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.739839 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.749827 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmmrh\" (UniqueName: \"kubernetes.io/projected/52e28a17-18b9-44a3-beec-cf688f768519-kube-api-access-gmmrh\") pod \"barbican-a6be-account-create-update-z9kh5\" (UID: \"52e28a17-18b9-44a3-beec-cf688f768519\") " pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.753133 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p977\" (UniqueName: \"kubernetes.io/projected/27f72293-3842-46ab-87ad-fe2345ea4b3a-kube-api-access-7p977\") pod \"neutron-c123-account-create-update-r29rm\" (UID: \"27f72293-3842-46ab-87ad-fe2345ea4b3a\") " pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.837761 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.853489 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.868966 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-67gcf"] Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:25.922512 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5253-account-create-update-zqf6n"] Mar 13 07:58:26 crc kubenswrapper[4876]: W0313 07:58:25.991046 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode16d2f43_d740_4edb_b41d_806e8bb28ff0.slice/crio-f9affec716aab49e6c5146cd8cf366bd5dbd54d163786796a6f9f5044ddd368e WatchSource:0}: Error finding container f9affec716aab49e6c5146cd8cf366bd5dbd54d163786796a6f9f5044ddd368e: Status 404 returned error can't find the container with id f9affec716aab49e6c5146cd8cf366bd5dbd54d163786796a6f9f5044ddd368e Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.347516 4876 generic.go:334] "Generic (PLEG): container finished" podID="0b06a84e-50b9-4e3f-b57e-4db1b023699d" containerID="79e044258014472c82a00a13c246a0d4373b7396c93188f36a1c7efe5d2fb7bd" exitCode=0 Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.347640 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-67gcf" event={"ID":"0b06a84e-50b9-4e3f-b57e-4db1b023699d","Type":"ContainerDied","Data":"79e044258014472c82a00a13c246a0d4373b7396c93188f36a1c7efe5d2fb7bd"} Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.347938 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-67gcf" event={"ID":"0b06a84e-50b9-4e3f-b57e-4db1b023699d","Type":"ContainerStarted","Data":"fe592599b8053706894507215a44b18fb79a777dc39ee1fbf84c163c1ab592cf"} Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.354589 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5253-account-create-update-zqf6n" event={"ID":"e16d2f43-d740-4edb-b41d-806e8bb28ff0","Type":"ContainerStarted","Data":"b85d91a2a04c2f2f36ad7d8b2e07b84b66502458e73bb82449707d5930ae404f"} Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.354632 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5253-account-create-update-zqf6n" event={"ID":"e16d2f43-d740-4edb-b41d-806e8bb28ff0","Type":"ContainerStarted","Data":"f9affec716aab49e6c5146cd8cf366bd5dbd54d163786796a6f9f5044ddd368e"} Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.391672 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-5253-account-create-update-zqf6n" podStartSLOduration=2.391641894 podStartE2EDuration="2.391641894s" podCreationTimestamp="2026-03-13 07:58:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:26.383844123 +0000 UTC m=+1166.054623385" watchObservedRunningTime="2026-03-13 07:58:26.391641894 +0000 UTC m=+1166.062420896" Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.767632 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jm9nf"] Mar 13 07:58:26 crc kubenswrapper[4876]: W0313 07:58:26.790960 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ccc93a4_d813_4a7e_965d_2fdce73b92b5.slice/crio-72e6fc2b130e773141433545c987dda0e5ca11e01247eead5d15aa02986d8b22 WatchSource:0}: Error finding container 72e6fc2b130e773141433545c987dda0e5ca11e01247eead5d15aa02986d8b22: Status 404 returned error can't find the container with id 72e6fc2b130e773141433545c987dda0e5ca11e01247eead5d15aa02986d8b22 Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.798851 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a6be-account-create-update-z9kh5"] Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.806333 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xwt9g"] Mar 13 07:58:26 crc kubenswrapper[4876]: W0313 07:58:26.806832 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52e28a17_18b9_44a3_beec_cf688f768519.slice/crio-730e49cc239687586107286cd7540e4ea4c5bf86fff84f03023f2a7c2969b771 WatchSource:0}: Error finding container 730e49cc239687586107286cd7540e4ea4c5bf86fff84f03023f2a7c2969b771: Status 404 returned error can't find the container with id 730e49cc239687586107286cd7540e4ea4c5bf86fff84f03023f2a7c2969b771 Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.814696 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-45rb2"] Mar 13 07:58:26 crc kubenswrapper[4876]: W0313 07:58:26.815133 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27f72293_3842_46ab_87ad_fe2345ea4b3a.slice/crio-24b93c8500a7ce8128b7e6989aeb4ea634a3bcb10c3a6099b17939e43103ab36 WatchSource:0}: Error finding container 24b93c8500a7ce8128b7e6989aeb4ea634a3bcb10c3a6099b17939e43103ab36: Status 404 returned error can't find the container with id 24b93c8500a7ce8128b7e6989aeb4ea634a3bcb10c3a6099b17939e43103ab36 Mar 13 07:58:26 crc kubenswrapper[4876]: I0313 07:58:26.820537 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c123-account-create-update-r29rm"] Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.391591 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jm9nf" event={"ID":"44f174ba-30f6-4440-9d1c-037d413b5f90","Type":"ContainerStarted","Data":"21004013435ebbea8d503cc46964a14804d066956a4b11aab90d7cba67ca1b1c"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.392076 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jm9nf" event={"ID":"44f174ba-30f6-4440-9d1c-037d413b5f90","Type":"ContainerStarted","Data":"a517369538cc55e18c54a2ad3d2bfb3eb60ae9b69457cf73f6b20cfa94aa3b52"} Mar 13 07:58:27 crc kubenswrapper[4876]: E0313 07:58:27.431050 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44f174ba_30f6_4440_9d1c_037d413b5f90.slice/crio-21004013435ebbea8d503cc46964a14804d066956a4b11aab90d7cba67ca1b1c.scope\": RecentStats: unable to find data in memory cache]" Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.432614 4876 generic.go:334] "Generic (PLEG): container finished" podID="e16d2f43-d740-4edb-b41d-806e8bb28ff0" containerID="b85d91a2a04c2f2f36ad7d8b2e07b84b66502458e73bb82449707d5930ae404f" exitCode=0 Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.432766 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5253-account-create-update-zqf6n" event={"ID":"e16d2f43-d740-4edb-b41d-806e8bb28ff0","Type":"ContainerDied","Data":"b85d91a2a04c2f2f36ad7d8b2e07b84b66502458e73bb82449707d5930ae404f"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.449553 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a6be-account-create-update-z9kh5" event={"ID":"52e28a17-18b9-44a3-beec-cf688f768519","Type":"ContainerStarted","Data":"9b8cd85de9248213a81da8270122abfd4dca68b5f7cb4c51879f0ae973ef3117"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.449634 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a6be-account-create-update-z9kh5" event={"ID":"52e28a17-18b9-44a3-beec-cf688f768519","Type":"ContainerStarted","Data":"730e49cc239687586107286cd7540e4ea4c5bf86fff84f03023f2a7c2969b771"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.457530 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-jm9nf" podStartSLOduration=2.457510772 podStartE2EDuration="2.457510772s" podCreationTimestamp="2026-03-13 07:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:27.423722262 +0000 UTC m=+1167.094501254" watchObservedRunningTime="2026-03-13 07:58:27.457510772 +0000 UTC m=+1167.128289754" Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.472756 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xwt9g" event={"ID":"4ccc93a4-d813-4a7e-965d-2fdce73b92b5","Type":"ContainerStarted","Data":"943035bd6fadc99ace78e87aa1087aa1f4c16b922ebe27ec8f1ff69c3e4d5a5e"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.472826 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xwt9g" event={"ID":"4ccc93a4-d813-4a7e-965d-2fdce73b92b5","Type":"ContainerStarted","Data":"72e6fc2b130e773141433545c987dda0e5ca11e01247eead5d15aa02986d8b22"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.483759 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-45rb2" event={"ID":"8523dd40-992c-4540-8bb0-2464da42a16c","Type":"ContainerStarted","Data":"6b730f04e0ab8ae5ded86163f8dd8356060ae7bbb6e349fe3498feed0193088d"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.505816 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-a6be-account-create-update-z9kh5" podStartSLOduration=2.5057926630000003 podStartE2EDuration="2.505792663s" podCreationTimestamp="2026-03-13 07:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:27.50213721 +0000 UTC m=+1167.172916202" watchObservedRunningTime="2026-03-13 07:58:27.505792663 +0000 UTC m=+1167.176571635" Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.521661 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c123-account-create-update-r29rm" event={"ID":"27f72293-3842-46ab-87ad-fe2345ea4b3a","Type":"ContainerStarted","Data":"6b137d93b7f9764555587381414a2363185119a97a017b13e0c0a831c8d99cd8"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.521945 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c123-account-create-update-r29rm" event={"ID":"27f72293-3842-46ab-87ad-fe2345ea4b3a","Type":"ContainerStarted","Data":"24b93c8500a7ce8128b7e6989aeb4ea634a3bcb10c3a6099b17939e43103ab36"} Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.574895 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c123-account-create-update-r29rm" podStartSLOduration=2.574858006 podStartE2EDuration="2.574858006s" podCreationTimestamp="2026-03-13 07:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:27.56233334 +0000 UTC m=+1167.233112322" watchObservedRunningTime="2026-03-13 07:58:27.574858006 +0000 UTC m=+1167.245636978" Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.575945 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-xwt9g" podStartSLOduration=2.575938386 podStartE2EDuration="2.575938386s" podCreationTimestamp="2026-03-13 07:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:27.537119264 +0000 UTC m=+1167.207898256" watchObservedRunningTime="2026-03-13 07:58:27.575938386 +0000 UTC m=+1167.246717368" Mar 13 07:58:27 crc kubenswrapper[4876]: I0313 07:58:27.946773 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.035434 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b06a84e-50b9-4e3f-b57e-4db1b023699d-operator-scripts\") pod \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\" (UID: \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\") " Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.035583 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swbrj\" (UniqueName: \"kubernetes.io/projected/0b06a84e-50b9-4e3f-b57e-4db1b023699d-kube-api-access-swbrj\") pod \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\" (UID: \"0b06a84e-50b9-4e3f-b57e-4db1b023699d\") " Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.037028 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b06a84e-50b9-4e3f-b57e-4db1b023699d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0b06a84e-50b9-4e3f-b57e-4db1b023699d" (UID: "0b06a84e-50b9-4e3f-b57e-4db1b023699d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.044635 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b06a84e-50b9-4e3f-b57e-4db1b023699d-kube-api-access-swbrj" (OuterVolumeSpecName: "kube-api-access-swbrj") pod "0b06a84e-50b9-4e3f-b57e-4db1b023699d" (UID: "0b06a84e-50b9-4e3f-b57e-4db1b023699d"). InnerVolumeSpecName "kube-api-access-swbrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.137917 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b06a84e-50b9-4e3f-b57e-4db1b023699d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.137963 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swbrj\" (UniqueName: \"kubernetes.io/projected/0b06a84e-50b9-4e3f-b57e-4db1b023699d-kube-api-access-swbrj\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.532185 4876 generic.go:334] "Generic (PLEG): container finished" podID="52e28a17-18b9-44a3-beec-cf688f768519" containerID="9b8cd85de9248213a81da8270122abfd4dca68b5f7cb4c51879f0ae973ef3117" exitCode=0 Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.532304 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a6be-account-create-update-z9kh5" event={"ID":"52e28a17-18b9-44a3-beec-cf688f768519","Type":"ContainerDied","Data":"9b8cd85de9248213a81da8270122abfd4dca68b5f7cb4c51879f0ae973ef3117"} Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.535951 4876 generic.go:334] "Generic (PLEG): container finished" podID="4ccc93a4-d813-4a7e-965d-2fdce73b92b5" containerID="943035bd6fadc99ace78e87aa1087aa1f4c16b922ebe27ec8f1ff69c3e4d5a5e" exitCode=0 Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.536085 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xwt9g" event={"ID":"4ccc93a4-d813-4a7e-965d-2fdce73b92b5","Type":"ContainerDied","Data":"943035bd6fadc99ace78e87aa1087aa1f4c16b922ebe27ec8f1ff69c3e4d5a5e"} Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.539790 4876 generic.go:334] "Generic (PLEG): container finished" podID="27f72293-3842-46ab-87ad-fe2345ea4b3a" containerID="6b137d93b7f9764555587381414a2363185119a97a017b13e0c0a831c8d99cd8" exitCode=0 Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.539858 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c123-account-create-update-r29rm" event={"ID":"27f72293-3842-46ab-87ad-fe2345ea4b3a","Type":"ContainerDied","Data":"6b137d93b7f9764555587381414a2363185119a97a017b13e0c0a831c8d99cd8"} Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.542721 4876 generic.go:334] "Generic (PLEG): container finished" podID="44f174ba-30f6-4440-9d1c-037d413b5f90" containerID="21004013435ebbea8d503cc46964a14804d066956a4b11aab90d7cba67ca1b1c" exitCode=0 Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.542816 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jm9nf" event={"ID":"44f174ba-30f6-4440-9d1c-037d413b5f90","Type":"ContainerDied","Data":"21004013435ebbea8d503cc46964a14804d066956a4b11aab90d7cba67ca1b1c"} Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.550048 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-67gcf" Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.550047 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-67gcf" event={"ID":"0b06a84e-50b9-4e3f-b57e-4db1b023699d","Type":"ContainerDied","Data":"fe592599b8053706894507215a44b18fb79a777dc39ee1fbf84c163c1ab592cf"} Mar 13 07:58:28 crc kubenswrapper[4876]: I0313 07:58:28.550168 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe592599b8053706894507215a44b18fb79a777dc39ee1fbf84c163c1ab592cf" Mar 13 07:58:31 crc kubenswrapper[4876]: I0313 07:58:31.952586 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.030688 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pc9zn"] Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.031164 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-pc9zn" podUID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" containerName="dnsmasq-dns" containerID="cri-o://63b8adbf89f8a0d05bacf662084557cbc5e86c6409efe944b5ea1eaec49f08e2" gracePeriod=10 Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.512165 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.519830 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.527124 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.551873 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.555836 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-operator-scripts\") pod \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\" (UID: \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.555909 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwbfl\" (UniqueName: \"kubernetes.io/projected/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-kube-api-access-gwbfl\") pod \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\" (UID: \"4ccc93a4-d813-4a7e-965d-2fdce73b92b5\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.557736 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ccc93a4-d813-4a7e-965d-2fdce73b92b5" (UID: "4ccc93a4-d813-4a7e-965d-2fdce73b92b5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.577633 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.591487 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-kube-api-access-gwbfl" (OuterVolumeSpecName: "kube-api-access-gwbfl") pod "4ccc93a4-d813-4a7e-965d-2fdce73b92b5" (UID: "4ccc93a4-d813-4a7e-965d-2fdce73b92b5"). InnerVolumeSpecName "kube-api-access-gwbfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.605103 4876 generic.go:334] "Generic (PLEG): container finished" podID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" containerID="63b8adbf89f8a0d05bacf662084557cbc5e86c6409efe944b5ea1eaec49f08e2" exitCode=0 Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.605321 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pc9zn" event={"ID":"7c2a1e76-01f4-4164-bdde-b7c30ff87058","Type":"ContainerDied","Data":"63b8adbf89f8a0d05bacf662084557cbc5e86c6409efe944b5ea1eaec49f08e2"} Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.609616 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a6be-account-create-update-z9kh5" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.609569 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a6be-account-create-update-z9kh5" event={"ID":"52e28a17-18b9-44a3-beec-cf688f768519","Type":"ContainerDied","Data":"730e49cc239687586107286cd7540e4ea4c5bf86fff84f03023f2a7c2969b771"} Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.609877 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="730e49cc239687586107286cd7540e4ea4c5bf86fff84f03023f2a7c2969b771" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.611515 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xwt9g" event={"ID":"4ccc93a4-d813-4a7e-965d-2fdce73b92b5","Type":"ContainerDied","Data":"72e6fc2b130e773141433545c987dda0e5ca11e01247eead5d15aa02986d8b22"} Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.611543 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72e6fc2b130e773141433545c987dda0e5ca11e01247eead5d15aa02986d8b22" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.611602 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xwt9g" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.615463 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c123-account-create-update-r29rm" event={"ID":"27f72293-3842-46ab-87ad-fe2345ea4b3a","Type":"ContainerDied","Data":"24b93c8500a7ce8128b7e6989aeb4ea634a3bcb10c3a6099b17939e43103ab36"} Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.615513 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24b93c8500a7ce8128b7e6989aeb4ea634a3bcb10c3a6099b17939e43103ab36" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.615609 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c123-account-create-update-r29rm" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.619138 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jm9nf" event={"ID":"44f174ba-30f6-4440-9d1c-037d413b5f90","Type":"ContainerDied","Data":"a517369538cc55e18c54a2ad3d2bfb3eb60ae9b69457cf73f6b20cfa94aa3b52"} Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.619166 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a517369538cc55e18c54a2ad3d2bfb3eb60ae9b69457cf73f6b20cfa94aa3b52" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.619230 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jm9nf" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.621200 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5253-account-create-update-zqf6n" event={"ID":"e16d2f43-d740-4edb-b41d-806e8bb28ff0","Type":"ContainerDied","Data":"f9affec716aab49e6c5146cd8cf366bd5dbd54d163786796a6f9f5044ddd368e"} Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.621221 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9affec716aab49e6c5146cd8cf366bd5dbd54d163786796a6f9f5044ddd368e" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.621283 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5253-account-create-update-zqf6n" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.657282 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p977\" (UniqueName: \"kubernetes.io/projected/27f72293-3842-46ab-87ad-fe2345ea4b3a-kube-api-access-7p977\") pod \"27f72293-3842-46ab-87ad-fe2345ea4b3a\" (UID: \"27f72293-3842-46ab-87ad-fe2345ea4b3a\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.657544 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e28a17-18b9-44a3-beec-cf688f768519-operator-scripts\") pod \"52e28a17-18b9-44a3-beec-cf688f768519\" (UID: \"52e28a17-18b9-44a3-beec-cf688f768519\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.657677 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f174ba-30f6-4440-9d1c-037d413b5f90-operator-scripts\") pod \"44f174ba-30f6-4440-9d1c-037d413b5f90\" (UID: \"44f174ba-30f6-4440-9d1c-037d413b5f90\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.657750 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmmrh\" (UniqueName: \"kubernetes.io/projected/52e28a17-18b9-44a3-beec-cf688f768519-kube-api-access-gmmrh\") pod \"52e28a17-18b9-44a3-beec-cf688f768519\" (UID: \"52e28a17-18b9-44a3-beec-cf688f768519\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.657817 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27f72293-3842-46ab-87ad-fe2345ea4b3a-operator-scripts\") pod \"27f72293-3842-46ab-87ad-fe2345ea4b3a\" (UID: \"27f72293-3842-46ab-87ad-fe2345ea4b3a\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.657903 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwczk\" (UniqueName: \"kubernetes.io/projected/e16d2f43-d740-4edb-b41d-806e8bb28ff0-kube-api-access-qwczk\") pod \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\" (UID: \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.658003 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf7kz\" (UniqueName: \"kubernetes.io/projected/44f174ba-30f6-4440-9d1c-037d413b5f90-kube-api-access-mf7kz\") pod \"44f174ba-30f6-4440-9d1c-037d413b5f90\" (UID: \"44f174ba-30f6-4440-9d1c-037d413b5f90\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.658161 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e16d2f43-d740-4edb-b41d-806e8bb28ff0-operator-scripts\") pod \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\" (UID: \"e16d2f43-d740-4edb-b41d-806e8bb28ff0\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.658751 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.658840 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwbfl\" (UniqueName: \"kubernetes.io/projected/4ccc93a4-d813-4a7e-965d-2fdce73b92b5-kube-api-access-gwbfl\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.660431 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e16d2f43-d740-4edb-b41d-806e8bb28ff0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e16d2f43-d740-4edb-b41d-806e8bb28ff0" (UID: "e16d2f43-d740-4edb-b41d-806e8bb28ff0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.660483 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52e28a17-18b9-44a3-beec-cf688f768519-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52e28a17-18b9-44a3-beec-cf688f768519" (UID: "52e28a17-18b9-44a3-beec-cf688f768519"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.660998 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44f174ba-30f6-4440-9d1c-037d413b5f90-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "44f174ba-30f6-4440-9d1c-037d413b5f90" (UID: "44f174ba-30f6-4440-9d1c-037d413b5f90"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.661387 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27f72293-3842-46ab-87ad-fe2345ea4b3a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "27f72293-3842-46ab-87ad-fe2345ea4b3a" (UID: "27f72293-3842-46ab-87ad-fe2345ea4b3a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.664269 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27f72293-3842-46ab-87ad-fe2345ea4b3a-kube-api-access-7p977" (OuterVolumeSpecName: "kube-api-access-7p977") pod "27f72293-3842-46ab-87ad-fe2345ea4b3a" (UID: "27f72293-3842-46ab-87ad-fe2345ea4b3a"). InnerVolumeSpecName "kube-api-access-7p977". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.666128 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e16d2f43-d740-4edb-b41d-806e8bb28ff0-kube-api-access-qwczk" (OuterVolumeSpecName: "kube-api-access-qwczk") pod "e16d2f43-d740-4edb-b41d-806e8bb28ff0" (UID: "e16d2f43-d740-4edb-b41d-806e8bb28ff0"). InnerVolumeSpecName "kube-api-access-qwczk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.666270 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52e28a17-18b9-44a3-beec-cf688f768519-kube-api-access-gmmrh" (OuterVolumeSpecName: "kube-api-access-gmmrh") pod "52e28a17-18b9-44a3-beec-cf688f768519" (UID: "52e28a17-18b9-44a3-beec-cf688f768519"). InnerVolumeSpecName "kube-api-access-gmmrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.693943 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44f174ba-30f6-4440-9d1c-037d413b5f90-kube-api-access-mf7kz" (OuterVolumeSpecName: "kube-api-access-mf7kz") pod "44f174ba-30f6-4440-9d1c-037d413b5f90" (UID: "44f174ba-30f6-4440-9d1c-037d413b5f90"). InnerVolumeSpecName "kube-api-access-mf7kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.759458 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.761869 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e16d2f43-d740-4edb-b41d-806e8bb28ff0-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.761909 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p977\" (UniqueName: \"kubernetes.io/projected/27f72293-3842-46ab-87ad-fe2345ea4b3a-kube-api-access-7p977\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.761922 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52e28a17-18b9-44a3-beec-cf688f768519-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.761935 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f174ba-30f6-4440-9d1c-037d413b5f90-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.761947 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmmrh\" (UniqueName: \"kubernetes.io/projected/52e28a17-18b9-44a3-beec-cf688f768519-kube-api-access-gmmrh\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.761956 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27f72293-3842-46ab-87ad-fe2345ea4b3a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.761966 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwczk\" (UniqueName: \"kubernetes.io/projected/e16d2f43-d740-4edb-b41d-806e8bb28ff0-kube-api-access-qwczk\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.761975 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf7kz\" (UniqueName: \"kubernetes.io/projected/44f174ba-30f6-4440-9d1c-037d413b5f90-kube-api-access-mf7kz\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.862834 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-nb\") pod \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.862914 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxlk8\" (UniqueName: \"kubernetes.io/projected/7c2a1e76-01f4-4164-bdde-b7c30ff87058-kube-api-access-mxlk8\") pod \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.862974 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-sb\") pod \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.863030 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-config\") pod \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.863071 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-dns-svc\") pod \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\" (UID: \"7c2a1e76-01f4-4164-bdde-b7c30ff87058\") " Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.867342 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c2a1e76-01f4-4164-bdde-b7c30ff87058-kube-api-access-mxlk8" (OuterVolumeSpecName: "kube-api-access-mxlk8") pod "7c2a1e76-01f4-4164-bdde-b7c30ff87058" (UID: "7c2a1e76-01f4-4164-bdde-b7c30ff87058"). InnerVolumeSpecName "kube-api-access-mxlk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.908892 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7c2a1e76-01f4-4164-bdde-b7c30ff87058" (UID: "7c2a1e76-01f4-4164-bdde-b7c30ff87058"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.908967 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7c2a1e76-01f4-4164-bdde-b7c30ff87058" (UID: "7c2a1e76-01f4-4164-bdde-b7c30ff87058"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.913082 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7c2a1e76-01f4-4164-bdde-b7c30ff87058" (UID: "7c2a1e76-01f4-4164-bdde-b7c30ff87058"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.914559 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-config" (OuterVolumeSpecName: "config") pod "7c2a1e76-01f4-4164-bdde-b7c30ff87058" (UID: "7c2a1e76-01f4-4164-bdde-b7c30ff87058"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.965859 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.965951 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.965971 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.965983 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c2a1e76-01f4-4164-bdde-b7c30ff87058-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:32 crc kubenswrapper[4876]: I0313 07:58:32.965999 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxlk8\" (UniqueName: \"kubernetes.io/projected/7c2a1e76-01f4-4164-bdde-b7c30ff87058-kube-api-access-mxlk8\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:33 crc kubenswrapper[4876]: I0313 07:58:33.641425 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-45rb2" event={"ID":"8523dd40-992c-4540-8bb0-2464da42a16c","Type":"ContainerStarted","Data":"bbe1a539a48c93a928b82c30e621bf5e872a6f21d864308820fdfefc31cecbc0"} Mar 13 07:58:33 crc kubenswrapper[4876]: I0313 07:58:33.647571 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pc9zn" event={"ID":"7c2a1e76-01f4-4164-bdde-b7c30ff87058","Type":"ContainerDied","Data":"91c2147e976802ffea7353a764902092d9ba57c0744581cd87415d9c68a41eee"} Mar 13 07:58:33 crc kubenswrapper[4876]: I0313 07:58:33.647647 4876 scope.go:117] "RemoveContainer" containerID="63b8adbf89f8a0d05bacf662084557cbc5e86c6409efe944b5ea1eaec49f08e2" Mar 13 07:58:33 crc kubenswrapper[4876]: I0313 07:58:33.647668 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-pc9zn" Mar 13 07:58:33 crc kubenswrapper[4876]: I0313 07:58:33.669231 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-45rb2" podStartSLOduration=2.19704962 podStartE2EDuration="8.669207368s" podCreationTimestamp="2026-03-13 07:58:25 +0000 UTC" firstStartedPulling="2026-03-13 07:58:26.83279457 +0000 UTC m=+1166.503573552" lastFinishedPulling="2026-03-13 07:58:33.304952318 +0000 UTC m=+1172.975731300" observedRunningTime="2026-03-13 07:58:33.65943765 +0000 UTC m=+1173.330216632" watchObservedRunningTime="2026-03-13 07:58:33.669207368 +0000 UTC m=+1173.339986350" Mar 13 07:58:33 crc kubenswrapper[4876]: I0313 07:58:33.672011 4876 scope.go:117] "RemoveContainer" containerID="5106a266925ccb985fb512f2a4c255255633daf36f8d7f87d07779a214bc2c34" Mar 13 07:58:33 crc kubenswrapper[4876]: I0313 07:58:33.681940 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pc9zn"] Mar 13 07:58:33 crc kubenswrapper[4876]: I0313 07:58:33.688967 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pc9zn"] Mar 13 07:58:35 crc kubenswrapper[4876]: I0313 07:58:35.049728 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" path="/var/lib/kubelet/pods/7c2a1e76-01f4-4164-bdde-b7c30ff87058/volumes" Mar 13 07:58:36 crc kubenswrapper[4876]: I0313 07:58:36.677769 4876 generic.go:334] "Generic (PLEG): container finished" podID="8523dd40-992c-4540-8bb0-2464da42a16c" containerID="bbe1a539a48c93a928b82c30e621bf5e872a6f21d864308820fdfefc31cecbc0" exitCode=0 Mar 13 07:58:36 crc kubenswrapper[4876]: I0313 07:58:36.677980 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-45rb2" event={"ID":"8523dd40-992c-4540-8bb0-2464da42a16c","Type":"ContainerDied","Data":"bbe1a539a48c93a928b82c30e621bf5e872a6f21d864308820fdfefc31cecbc0"} Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.037254 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.067068 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-config-data\") pod \"8523dd40-992c-4540-8bb0-2464da42a16c\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.067181 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-combined-ca-bundle\") pod \"8523dd40-992c-4540-8bb0-2464da42a16c\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.067218 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbh6k\" (UniqueName: \"kubernetes.io/projected/8523dd40-992c-4540-8bb0-2464da42a16c-kube-api-access-qbh6k\") pod \"8523dd40-992c-4540-8bb0-2464da42a16c\" (UID: \"8523dd40-992c-4540-8bb0-2464da42a16c\") " Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.074605 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8523dd40-992c-4540-8bb0-2464da42a16c-kube-api-access-qbh6k" (OuterVolumeSpecName: "kube-api-access-qbh6k") pod "8523dd40-992c-4540-8bb0-2464da42a16c" (UID: "8523dd40-992c-4540-8bb0-2464da42a16c"). InnerVolumeSpecName "kube-api-access-qbh6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.093604 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8523dd40-992c-4540-8bb0-2464da42a16c" (UID: "8523dd40-992c-4540-8bb0-2464da42a16c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.114447 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-config-data" (OuterVolumeSpecName: "config-data") pod "8523dd40-992c-4540-8bb0-2464da42a16c" (UID: "8523dd40-992c-4540-8bb0-2464da42a16c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.169960 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.170024 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8523dd40-992c-4540-8bb0-2464da42a16c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.170051 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbh6k\" (UniqueName: \"kubernetes.io/projected/8523dd40-992c-4540-8bb0-2464da42a16c-kube-api-access-qbh6k\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.697128 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-45rb2" event={"ID":"8523dd40-992c-4540-8bb0-2464da42a16c","Type":"ContainerDied","Data":"6b730f04e0ab8ae5ded86163f8dd8356060ae7bbb6e349fe3498feed0193088d"} Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.697180 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b730f04e0ab8ae5ded86163f8dd8356060ae7bbb6e349fe3498feed0193088d" Mar 13 07:58:38 crc kubenswrapper[4876]: I0313 07:58:38.697337 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-45rb2" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.027907 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6hbtk"] Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028411 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8523dd40-992c-4540-8bb0-2464da42a16c" containerName="keystone-db-sync" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028436 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8523dd40-992c-4540-8bb0-2464da42a16c" containerName="keystone-db-sync" Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028452 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e16d2f43-d740-4edb-b41d-806e8bb28ff0" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028459 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e16d2f43-d740-4edb-b41d-806e8bb28ff0" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028478 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f72293-3842-46ab-87ad-fe2345ea4b3a" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028486 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f72293-3842-46ab-87ad-fe2345ea4b3a" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028504 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccc93a4-d813-4a7e-965d-2fdce73b92b5" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028512 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccc93a4-d813-4a7e-965d-2fdce73b92b5" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028525 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44f174ba-30f6-4440-9d1c-037d413b5f90" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028531 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44f174ba-30f6-4440-9d1c-037d413b5f90" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028545 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" containerName="init" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028553 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" containerName="init" Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028566 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b06a84e-50b9-4e3f-b57e-4db1b023699d" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028573 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b06a84e-50b9-4e3f-b57e-4db1b023699d" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028590 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" containerName="dnsmasq-dns" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028598 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" containerName="dnsmasq-dns" Mar 13 07:58:39 crc kubenswrapper[4876]: E0313 07:58:39.028613 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e28a17-18b9-44a3-beec-cf688f768519" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028622 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e28a17-18b9-44a3-beec-cf688f768519" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028846 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b06a84e-50b9-4e3f-b57e-4db1b023699d" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028865 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="52e28a17-18b9-44a3-beec-cf688f768519" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028883 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e16d2f43-d740-4edb-b41d-806e8bb28ff0" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028893 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ccc93a4-d813-4a7e-965d-2fdce73b92b5" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028906 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="44f174ba-30f6-4440-9d1c-037d413b5f90" containerName="mariadb-database-create" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028914 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f72293-3842-46ab-87ad-fe2345ea4b3a" containerName="mariadb-account-create-update" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028926 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c2a1e76-01f4-4164-bdde-b7c30ff87058" containerName="dnsmasq-dns" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.028942 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8523dd40-992c-4540-8bb0-2464da42a16c" containerName="keystone-db-sync" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.030083 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.050784 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-j4z2h"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.052013 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.055695 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.055768 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.056208 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.063109 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.064591 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-btt8h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.070472 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-j4z2h"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.141994 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-config-data\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.142080 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-fernet-keys\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.142223 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-scripts\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.142411 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-credential-keys\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.142450 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-combined-ca-bundle\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.142515 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5v6b\" (UniqueName: \"kubernetes.io/projected/95f47410-f357-47ce-9f29-f6948f02ed45-kube-api-access-n5v6b\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.213149 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6hbtk"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.243801 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-config-data\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.243863 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-config\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.243897 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-fernet-keys\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.243929 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.243954 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.243988 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-scripts\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.244050 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-credential-keys\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.244082 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-combined-ca-bundle\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.244124 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5v6b\" (UniqueName: \"kubernetes.io/projected/95f47410-f357-47ce-9f29-f6948f02ed45-kube-api-access-n5v6b\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.244163 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mp62\" (UniqueName: \"kubernetes.io/projected/b81d3cbc-3656-4207-adad-46b11ed67d92-kube-api-access-6mp62\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.244187 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.244226 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.252281 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-fernet-keys\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.252979 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-scripts\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.253837 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-credential-keys\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.256091 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-combined-ca-bundle\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.275167 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-config-data\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.294880 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5v6b\" (UniqueName: \"kubernetes.io/projected/95f47410-f357-47ce-9f29-f6948f02ed45-kube-api-access-n5v6b\") pod \"keystone-bootstrap-j4z2h\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.318491 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-59996cc657-lxnh9"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.320298 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.322542 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.327683 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.328077 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-vgf6v" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.328214 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345663 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-horizon-secret-key\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345730 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mp62\" (UniqueName: \"kubernetes.io/projected/b81d3cbc-3656-4207-adad-46b11ed67d92-kube-api-access-6mp62\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345751 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345778 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345798 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-config-data\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345818 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xntrc\" (UniqueName: \"kubernetes.io/projected/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-kube-api-access-xntrc\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345842 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-config\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345868 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345885 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-logs\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345911 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.345939 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-scripts\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.347009 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.347173 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.347843 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.348898 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.350164 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-config\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.350348 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59996cc657-lxnh9"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.370977 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.380144 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-ltwdh"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.383289 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.385694 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.386046 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5gdvr" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.386721 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.399117 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mp62\" (UniqueName: \"kubernetes.io/projected/b81d3cbc-3656-4207-adad-46b11ed67d92-kube-api-access-6mp62\") pod \"dnsmasq-dns-847c4cc679-6hbtk\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.431730 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-ltwdh"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.447774 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-config-data\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.447831 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xntrc\" (UniqueName: \"kubernetes.io/projected/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-kube-api-access-xntrc\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.447871 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-logs\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.447904 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-scripts\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.447946 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-horizon-secret-key\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.453832 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-horizon-secret-key\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.454145 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-logs\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.458201 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-scripts\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.463655 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-config-data\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.554529 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz79q\" (UniqueName: \"kubernetes.io/projected/5d42486d-85b1-4384-91fc-10552a2c404b-kube-api-access-sz79q\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.554591 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-config-data\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.554624 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-db-sync-config-data\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.554652 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d42486d-85b1-4384-91fc-10552a2c404b-etc-machine-id\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.554687 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-scripts\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.554711 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-combined-ca-bundle\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.559206 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-hsm7j"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.560484 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.577971 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.584025 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wr4mz" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.588177 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.633679 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hsm7j"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.651715 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.659251 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz79q\" (UniqueName: \"kubernetes.io/projected/5d42486d-85b1-4384-91fc-10552a2c404b-kube-api-access-sz79q\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.659332 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-config-data\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.659376 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-db-sync-config-data\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.659413 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d42486d-85b1-4384-91fc-10552a2c404b-etc-machine-id\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.659455 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-scripts\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.659480 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-combined-ca-bundle\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.660374 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d42486d-85b1-4384-91fc-10552a2c404b-etc-machine-id\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.668597 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-combined-ca-bundle\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.671529 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-config-data\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.671805 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-db-sync-config-data\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.673801 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-scripts\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.698567 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-l75wd"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.699922 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.709331 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n5vvq" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.709477 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.717903 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6hbtk"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.718807 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.742087 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz79q\" (UniqueName: \"kubernetes.io/projected/5d42486d-85b1-4384-91fc-10552a2c404b-kube-api-access-sz79q\") pod \"cinder-db-sync-ltwdh\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.757412 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-l75wd"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.781336 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-combined-ca-bundle\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.781414 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-config\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.781465 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27g6x\" (UniqueName: \"kubernetes.io/projected/669de0e1-d1a3-4298-bee2-e3713fd7f652-kube-api-access-27g6x\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.781529 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-combined-ca-bundle\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.781612 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-config-data\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.781873 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7scb8\" (UniqueName: \"kubernetes.io/projected/e0b8464c-522b-44cc-8943-0f0770c2310b-kube-api-access-7scb8\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.781897 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/669de0e1-d1a3-4298-bee2-e3713fd7f652-logs\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.782026 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-scripts\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.796445 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-654b7b977c-7x6lq"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.797851 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xntrc\" (UniqueName: \"kubernetes.io/projected/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-kube-api-access-xntrc\") pod \"horizon-59996cc657-lxnh9\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.797975 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.824322 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-rzbxc"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.825629 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.837856 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-p54ct" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.838104 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.838114 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.845620 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.858863 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.860488 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.869052 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.869212 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.869834 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rzbxc"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.883826 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27g6x\" (UniqueName: \"kubernetes.io/projected/669de0e1-d1a3-4298-bee2-e3713fd7f652-kube-api-access-27g6x\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.883876 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-combined-ca-bundle\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.883912 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-config-data\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.884067 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7scb8\" (UniqueName: \"kubernetes.io/projected/e0b8464c-522b-44cc-8943-0f0770c2310b-kube-api-access-7scb8\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.884086 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/669de0e1-d1a3-4298-bee2-e3713fd7f652-logs\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.884134 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-scripts\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.884157 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-combined-ca-bundle\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.884177 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-config\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.884292 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-654b7b977c-7x6lq"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.888641 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/669de0e1-d1a3-4298-bee2-e3713fd7f652-logs\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.902551 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-scripts\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.903609 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-combined-ca-bundle\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.905824 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-config\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.908134 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-combined-ca-bundle\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.919293 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5v9sw"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.933026 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.940935 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5v9sw"] Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.961858 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27g6x\" (UniqueName: \"kubernetes.io/projected/669de0e1-d1a3-4298-bee2-e3713fd7f652-kube-api-access-27g6x\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.968748 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.983755 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-config-data\") pod \"placement-db-sync-l75wd\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.984924 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7scb8\" (UniqueName: \"kubernetes.io/projected/e0b8464c-522b-44cc-8943-0f0770c2310b-kube-api-access-7scb8\") pod \"neutron-db-sync-hsm7j\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986070 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-logs\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986186 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92grg\" (UniqueName: \"kubernetes.io/projected/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-kube-api-access-92grg\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986282 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986357 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-config-data\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986416 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-scripts\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986504 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-combined-ca-bundle\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986694 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-config-data\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986790 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-horizon-secret-key\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986865 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-log-httpd\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986899 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-run-httpd\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986933 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-scripts\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986971 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tndxt\" (UniqueName: \"kubernetes.io/projected/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-kube-api-access-tndxt\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.986998 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-db-sync-config-data\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.987072 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:39 crc kubenswrapper[4876]: I0313 07:58:39.987135 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb8qj\" (UniqueName: \"kubernetes.io/projected/c287c63f-40c2-427d-a58e-fea0297c947b-kube-api-access-sb8qj\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.057909 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-l75wd" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093062 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-log-httpd\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093147 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqt2q\" (UniqueName: \"kubernetes.io/projected/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-kube-api-access-pqt2q\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093175 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-run-httpd\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093205 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-scripts\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093251 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tndxt\" (UniqueName: \"kubernetes.io/projected/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-kube-api-access-tndxt\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093281 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-db-sync-config-data\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093319 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093348 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb8qj\" (UniqueName: \"kubernetes.io/projected/c287c63f-40c2-427d-a58e-fea0297c947b-kube-api-access-sb8qj\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093381 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-logs\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093414 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92grg\" (UniqueName: \"kubernetes.io/projected/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-kube-api-access-92grg\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093449 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093477 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093501 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093525 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-config-data\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093549 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-scripts\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093581 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093632 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-combined-ca-bundle\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093675 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-config-data\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093701 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093738 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-horizon-secret-key\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.093763 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-config\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.095948 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-log-httpd\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.096375 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-run-httpd\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.097965 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-scripts\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.098780 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-logs\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.103094 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-config-data\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.108668 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-scripts\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.109936 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.110179 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.112358 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-combined-ca-bundle\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.113065 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-config-data\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.113876 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-horizon-secret-key\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.115025 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-db-sync-config-data\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.184075 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.190195 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.193586 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xxnmv" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.193617 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.193625 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.194227 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.197192 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.197287 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.197362 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.197473 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.197541 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-config\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.197609 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqt2q\" (UniqueName: \"kubernetes.io/projected/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-kube-api-access-pqt2q\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.198494 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.199649 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.199756 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.200473 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.200595 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-config\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.200979 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.210931 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.232570 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqt2q\" (UniqueName: \"kubernetes.io/projected/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-kube-api-access-pqt2q\") pod \"dnsmasq-dns-785d8bcb8c-5v9sw\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.232751 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tndxt\" (UniqueName: \"kubernetes.io/projected/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-kube-api-access-tndxt\") pod \"barbican-db-sync-rzbxc\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.233184 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92grg\" (UniqueName: \"kubernetes.io/projected/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-kube-api-access-92grg\") pod \"horizon-654b7b977c-7x6lq\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.234471 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb8qj\" (UniqueName: \"kubernetes.io/projected/c287c63f-40c2-427d-a58e-fea0297c947b-kube-api-access-sb8qj\") pod \"ceilometer-0\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.263342 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.300997 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.301056 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.301092 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.301119 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-logs\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.301139 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.306437 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.306965 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t24vb\" (UniqueName: \"kubernetes.io/projected/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-kube-api-access-t24vb\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.306999 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.325317 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.409914 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.409978 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.410013 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.410048 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-logs\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.410073 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.410117 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.410149 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.410166 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t24vb\" (UniqueName: \"kubernetes.io/projected/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-kube-api-access-t24vb\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.412774 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-logs\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.417002 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.419803 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.422910 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.428091 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.433065 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.435003 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.464229 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.466440 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.488082 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.488818 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.496742 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.497142 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.505627 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.506045 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.508764 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t24vb\" (UniqueName: \"kubernetes.io/projected/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-kube-api-access-t24vb\") pod \"glance-default-external-api-0\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.517188 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.615937 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.616477 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.616502 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.616566 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nrb7\" (UniqueName: \"kubernetes.io/projected/42968848-d1aa-4140-9a19-0218cdbc98d8-kube-api-access-5nrb7\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.616614 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.616662 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.616683 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.616854 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.638521 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-ltwdh"] Mar 13 07:58:40 crc kubenswrapper[4876]: W0313 07:58:40.658392 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d42486d_85b1_4384_91fc_10552a2c404b.slice/crio-98b1fb16c85acd3567987e5ddc33472b976ee850ad6261f8278fbb918351bcbf WatchSource:0}: Error finding container 98b1fb16c85acd3567987e5ddc33472b976ee850ad6261f8278fbb918351bcbf: Status 404 returned error can't find the container with id 98b1fb16c85acd3567987e5ddc33472b976ee850ad6261f8278fbb918351bcbf Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.678287 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-j4z2h"] Mar 13 07:58:40 crc kubenswrapper[4876]: W0313 07:58:40.683664 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95f47410_f357_47ce_9f29_f6948f02ed45.slice/crio-27b3ec3de9e26dd4eed47461e95f16621920baf010000cd68ed4a3a9c496e178 WatchSource:0}: Error finding container 27b3ec3de9e26dd4eed47461e95f16621920baf010000cd68ed4a3a9c496e178: Status 404 returned error can't find the container with id 27b3ec3de9e26dd4eed47461e95f16621920baf010000cd68ed4a3a9c496e178 Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.720185 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.720220 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.720278 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nrb7\" (UniqueName: \"kubernetes.io/projected/42968848-d1aa-4140-9a19-0218cdbc98d8-kube-api-access-5nrb7\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.720310 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.720348 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.720366 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.720394 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.720432 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.722398 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.722659 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.722741 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.725410 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.727092 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.730208 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.744028 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.758788 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nrb7\" (UniqueName: \"kubernetes.io/projected/42968848-d1aa-4140-9a19-0218cdbc98d8-kube-api-access-5nrb7\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.765971 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ltwdh" event={"ID":"5d42486d-85b1-4384-91fc-10552a2c404b","Type":"ContainerStarted","Data":"98b1fb16c85acd3567987e5ddc33472b976ee850ad6261f8278fbb918351bcbf"} Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.770923 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.771882 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-j4z2h" event={"ID":"95f47410-f357-47ce-9f29-f6948f02ed45","Type":"ContainerStarted","Data":"27b3ec3de9e26dd4eed47461e95f16621920baf010000cd68ed4a3a9c496e178"} Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.833925 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 07:58:40 crc kubenswrapper[4876]: I0313 07:58:40.908976 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6hbtk"] Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.102563 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-l75wd"] Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.102601 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.102611 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hsm7j"] Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.106535 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59996cc657-lxnh9"] Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.246256 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-654b7b977c-7x6lq"] Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.262811 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5v9sw"] Mar 13 07:58:41 crc kubenswrapper[4876]: W0313 07:58:41.263824 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b6f7e51_08c6_4e1a_b099_31ffc41866ea.slice/crio-8f694ac305e489193cf2a6a654ae1eeefb1dcd174dca1dec02ce1c8ec9d37098 WatchSource:0}: Error finding container 8f694ac305e489193cf2a6a654ae1eeefb1dcd174dca1dec02ce1c8ec9d37098: Status 404 returned error can't find the container with id 8f694ac305e489193cf2a6a654ae1eeefb1dcd174dca1dec02ce1c8ec9d37098 Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.389650 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rzbxc"] Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.785603 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.843718 4876 generic.go:334] "Generic (PLEG): container finished" podID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerID="e6471cb3f5e9fbe3c96869a454d8b4b0fedc8deb7f34dc47c585acb5bb24e3f8" exitCode=0 Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.843829 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" event={"ID":"2b6f7e51-08c6-4e1a-b099-31ffc41866ea","Type":"ContainerDied","Data":"e6471cb3f5e9fbe3c96869a454d8b4b0fedc8deb7f34dc47c585acb5bb24e3f8"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.843874 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" event={"ID":"2b6f7e51-08c6-4e1a-b099-31ffc41866ea","Type":"ContainerStarted","Data":"8f694ac305e489193cf2a6a654ae1eeefb1dcd174dca1dec02ce1c8ec9d37098"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.849809 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-l75wd" event={"ID":"669de0e1-d1a3-4298-bee2-e3713fd7f652","Type":"ContainerStarted","Data":"fc5a33cf3379b1ddc5a013724c0ef1ab861099c497d927f6d741f0298b6570d3"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.863652 4876 generic.go:334] "Generic (PLEG): container finished" podID="b81d3cbc-3656-4207-adad-46b11ed67d92" containerID="d3dc84e4762f2edb152dc30445ca31aedb292244ad494f557eb2f6919ec7694f" exitCode=0 Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.863745 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" event={"ID":"b81d3cbc-3656-4207-adad-46b11ed67d92","Type":"ContainerDied","Data":"d3dc84e4762f2edb152dc30445ca31aedb292244ad494f557eb2f6919ec7694f"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.863784 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" event={"ID":"b81d3cbc-3656-4207-adad-46b11ed67d92","Type":"ContainerStarted","Data":"01dd4b02c2f105f642173fa9cbd1bf8c5613608a9fd3926f635a1c7c4ed874aa"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.866920 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59996cc657-lxnh9" event={"ID":"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a","Type":"ContainerStarted","Data":"4a3147076ee9e24ee442e33b726dc0146d6b7647b44bb3da55a3e46d011ede22"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.869556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rzbxc" event={"ID":"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32","Type":"ContainerStarted","Data":"9cec9de6c021b4153884f9cca0464e4b3e939357dc84f7d753dec6c76994a476"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.929819 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hsm7j" event={"ID":"e0b8464c-522b-44cc-8943-0f0770c2310b","Type":"ContainerStarted","Data":"6c9ccd7f9ccc1773a10fad2983c6fae777e122684bccaaa3d24a91f678f5ce2a"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.930517 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hsm7j" event={"ID":"e0b8464c-522b-44cc-8943-0f0770c2310b","Type":"ContainerStarted","Data":"a72baf22ee60aadf4f309a3915fc3a280afe39d744b04ca045c6f063d397862c"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.933457 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerStarted","Data":"d896f9559f6684b16d084597175cba274598b862a29dcbcddd07b3232f18d142"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.935127 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-j4z2h" event={"ID":"95f47410-f357-47ce-9f29-f6948f02ed45","Type":"ContainerStarted","Data":"d920151cf960cd9011266a42d377ced5f63cc0564624c236cf81e9406ae2903b"} Mar 13 07:58:41 crc kubenswrapper[4876]: I0313 07:58:41.961597 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b7b977c-7x6lq" event={"ID":"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c","Type":"ContainerStarted","Data":"69069459a42e0807683906e45d5ffff6fa3ff077a928cb5beb561d8c276ae755"} Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.020038 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-j4z2h" podStartSLOduration=4.020013598 podStartE2EDuration="4.020013598s" podCreationTimestamp="2026-03-13 07:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:42.019619727 +0000 UTC m=+1181.690398709" watchObservedRunningTime="2026-03-13 07:58:42.020013598 +0000 UTC m=+1181.690792580" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.040473 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-hsm7j" podStartSLOduration=3.040431729 podStartE2EDuration="3.040431729s" podCreationTimestamp="2026-03-13 07:58:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:41.972319073 +0000 UTC m=+1181.643098055" watchObservedRunningTime="2026-03-13 07:58:42.040431729 +0000 UTC m=+1181.711210731" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.130580 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.144155 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59996cc657-lxnh9"] Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.203758 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5df7599fc5-4djfw"] Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.205957 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.243063 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5df7599fc5-4djfw"] Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.284164 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxfxt\" (UniqueName: \"kubernetes.io/projected/210baf38-0b27-4268-a313-fc13d56f55e1-kube-api-access-lxfxt\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.284416 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-config-data\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.284580 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/210baf38-0b27-4268-a313-fc13d56f55e1-horizon-secret-key\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.284617 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/210baf38-0b27-4268-a313-fc13d56f55e1-logs\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.284757 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-scripts\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.301883 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.362764 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.376696 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.386518 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-config-data\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.386601 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/210baf38-0b27-4268-a313-fc13d56f55e1-horizon-secret-key\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.386625 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/210baf38-0b27-4268-a313-fc13d56f55e1-logs\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.386666 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-scripts\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.386740 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxfxt\" (UniqueName: \"kubernetes.io/projected/210baf38-0b27-4268-a313-fc13d56f55e1-kube-api-access-lxfxt\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.387448 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/210baf38-0b27-4268-a313-fc13d56f55e1-logs\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.387933 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-scripts\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.388313 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-config-data\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.391974 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/210baf38-0b27-4268-a313-fc13d56f55e1-horizon-secret-key\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.418084 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxfxt\" (UniqueName: \"kubernetes.io/projected/210baf38-0b27-4268-a313-fc13d56f55e1-kube-api-access-lxfxt\") pod \"horizon-5df7599fc5-4djfw\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.558177 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.577921 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.692365 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-swift-storage-0\") pod \"b81d3cbc-3656-4207-adad-46b11ed67d92\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.692578 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-sb\") pod \"b81d3cbc-3656-4207-adad-46b11ed67d92\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.692624 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-config\") pod \"b81d3cbc-3656-4207-adad-46b11ed67d92\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.692700 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mp62\" (UniqueName: \"kubernetes.io/projected/b81d3cbc-3656-4207-adad-46b11ed67d92-kube-api-access-6mp62\") pod \"b81d3cbc-3656-4207-adad-46b11ed67d92\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.692726 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-svc\") pod \"b81d3cbc-3656-4207-adad-46b11ed67d92\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.692758 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-nb\") pod \"b81d3cbc-3656-4207-adad-46b11ed67d92\" (UID: \"b81d3cbc-3656-4207-adad-46b11ed67d92\") " Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.706512 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b81d3cbc-3656-4207-adad-46b11ed67d92-kube-api-access-6mp62" (OuterVolumeSpecName: "kube-api-access-6mp62") pod "b81d3cbc-3656-4207-adad-46b11ed67d92" (UID: "b81d3cbc-3656-4207-adad-46b11ed67d92"). InnerVolumeSpecName "kube-api-access-6mp62". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.745413 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b81d3cbc-3656-4207-adad-46b11ed67d92" (UID: "b81d3cbc-3656-4207-adad-46b11ed67d92"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.748798 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b81d3cbc-3656-4207-adad-46b11ed67d92" (UID: "b81d3cbc-3656-4207-adad-46b11ed67d92"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.748977 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-config" (OuterVolumeSpecName: "config") pod "b81d3cbc-3656-4207-adad-46b11ed67d92" (UID: "b81d3cbc-3656-4207-adad-46b11ed67d92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.765103 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b81d3cbc-3656-4207-adad-46b11ed67d92" (UID: "b81d3cbc-3656-4207-adad-46b11ed67d92"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.792513 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b81d3cbc-3656-4207-adad-46b11ed67d92" (UID: "b81d3cbc-3656-4207-adad-46b11ed67d92"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.798324 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.798363 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.798374 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.798385 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mp62\" (UniqueName: \"kubernetes.io/projected/b81d3cbc-3656-4207-adad-46b11ed67d92-kube-api-access-6mp62\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.798396 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.798404 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b81d3cbc-3656-4207-adad-46b11ed67d92-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.988921 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"42968848-d1aa-4140-9a19-0218cdbc98d8","Type":"ContainerStarted","Data":"0c6749d256b69aa33d06ffeb3ec1ee909b5a65ded73db1cd690e68222b2c69c0"} Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.991920 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34","Type":"ContainerStarted","Data":"407adf5d98657b8252d8abcf1f1b56219361c33bb87e3a4829c978690e98b234"} Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.998962 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" event={"ID":"2b6f7e51-08c6-4e1a-b099-31ffc41866ea","Type":"ContainerStarted","Data":"36fff9f513672061d505049d00b22d172bdafe59341929dfa7826153aec5016e"} Mar 13 07:58:42 crc kubenswrapper[4876]: I0313 07:58:42.999174 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:43 crc kubenswrapper[4876]: I0313 07:58:43.006172 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" event={"ID":"b81d3cbc-3656-4207-adad-46b11ed67d92","Type":"ContainerDied","Data":"01dd4b02c2f105f642173fa9cbd1bf8c5613608a9fd3926f635a1c7c4ed874aa"} Mar 13 07:58:43 crc kubenswrapper[4876]: I0313 07:58:43.006220 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6hbtk" Mar 13 07:58:43 crc kubenswrapper[4876]: I0313 07:58:43.006248 4876 scope.go:117] "RemoveContainer" containerID="d3dc84e4762f2edb152dc30445ca31aedb292244ad494f557eb2f6919ec7694f" Mar 13 07:58:43 crc kubenswrapper[4876]: I0313 07:58:43.030339 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" podStartSLOduration=4.030320707 podStartE2EDuration="4.030320707s" podCreationTimestamp="2026-03-13 07:58:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:43.02689501 +0000 UTC m=+1182.697673992" watchObservedRunningTime="2026-03-13 07:58:43.030320707 +0000 UTC m=+1182.701099689" Mar 13 07:58:43 crc kubenswrapper[4876]: I0313 07:58:43.094696 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6hbtk"] Mar 13 07:58:43 crc kubenswrapper[4876]: I0313 07:58:43.112946 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6hbtk"] Mar 13 07:58:43 crc kubenswrapper[4876]: I0313 07:58:43.127895 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5df7599fc5-4djfw"] Mar 13 07:58:44 crc kubenswrapper[4876]: I0313 07:58:44.038460 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"42968848-d1aa-4140-9a19-0218cdbc98d8","Type":"ContainerStarted","Data":"43eb9beea83cea81bfb0030fd545e7ecb5558c5a0ceb54835ff5e8cdaa8bfae4"} Mar 13 07:58:44 crc kubenswrapper[4876]: I0313 07:58:44.043663 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5df7599fc5-4djfw" event={"ID":"210baf38-0b27-4268-a313-fc13d56f55e1","Type":"ContainerStarted","Data":"7f1b2d62ae4e0c155abced5046679754bc23661e9124c16d9d09f846e123019b"} Mar 13 07:58:44 crc kubenswrapper[4876]: I0313 07:58:44.053924 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34","Type":"ContainerStarted","Data":"9ba0d1199b1516989333cfea63c6c3fc6a122aac7a1511c4383a6dc5207a4067"} Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.048229 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b81d3cbc-3656-4207-adad-46b11ed67d92" path="/var/lib/kubelet/pods/b81d3cbc-3656-4207-adad-46b11ed67d92/volumes" Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.113925 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34","Type":"ContainerStarted","Data":"89d2bc28ca15c6767b57fe40afe46636199325610ce16c3bbab01bd019fdcc49"} Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.114121 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerName="glance-log" containerID="cri-o://9ba0d1199b1516989333cfea63c6c3fc6a122aac7a1511c4383a6dc5207a4067" gracePeriod=30 Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.114254 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerName="glance-httpd" containerID="cri-o://89d2bc28ca15c6767b57fe40afe46636199325610ce16c3bbab01bd019fdcc49" gracePeriod=30 Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.124229 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"42968848-d1aa-4140-9a19-0218cdbc98d8","Type":"ContainerStarted","Data":"3623ef20ce0509f3a529da1f907a34545f69dea1c16c361e96d6ee4a558c49c3"} Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.124576 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerName="glance-httpd" containerID="cri-o://3623ef20ce0509f3a529da1f907a34545f69dea1c16c361e96d6ee4a558c49c3" gracePeriod=30 Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.124559 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerName="glance-log" containerID="cri-o://43eb9beea83cea81bfb0030fd545e7ecb5558c5a0ceb54835ff5e8cdaa8bfae4" gracePeriod=30 Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.153682 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.153659352 podStartE2EDuration="5.153659352s" podCreationTimestamp="2026-03-13 07:58:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:45.144887543 +0000 UTC m=+1184.815666535" watchObservedRunningTime="2026-03-13 07:58:45.153659352 +0000 UTC m=+1184.824438404" Mar 13 07:58:45 crc kubenswrapper[4876]: I0313 07:58:45.204318 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.204295441 podStartE2EDuration="6.204295441s" podCreationTimestamp="2026-03-13 07:58:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:58:45.185521867 +0000 UTC m=+1184.856300849" watchObservedRunningTime="2026-03-13 07:58:45.204295441 +0000 UTC m=+1184.875074423" Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.142204 4876 generic.go:334] "Generic (PLEG): container finished" podID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerID="3623ef20ce0509f3a529da1f907a34545f69dea1c16c361e96d6ee4a558c49c3" exitCode=0 Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.142631 4876 generic.go:334] "Generic (PLEG): container finished" podID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerID="43eb9beea83cea81bfb0030fd545e7ecb5558c5a0ceb54835ff5e8cdaa8bfae4" exitCode=143 Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.142340 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"42968848-d1aa-4140-9a19-0218cdbc98d8","Type":"ContainerDied","Data":"3623ef20ce0509f3a529da1f907a34545f69dea1c16c361e96d6ee4a558c49c3"} Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.142759 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"42968848-d1aa-4140-9a19-0218cdbc98d8","Type":"ContainerDied","Data":"43eb9beea83cea81bfb0030fd545e7ecb5558c5a0ceb54835ff5e8cdaa8bfae4"} Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.148567 4876 generic.go:334] "Generic (PLEG): container finished" podID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerID="89d2bc28ca15c6767b57fe40afe46636199325610ce16c3bbab01bd019fdcc49" exitCode=143 Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.148609 4876 generic.go:334] "Generic (PLEG): container finished" podID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerID="9ba0d1199b1516989333cfea63c6c3fc6a122aac7a1511c4383a6dc5207a4067" exitCode=143 Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.148742 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34","Type":"ContainerDied","Data":"89d2bc28ca15c6767b57fe40afe46636199325610ce16c3bbab01bd019fdcc49"} Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.148779 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34","Type":"ContainerDied","Data":"9ba0d1199b1516989333cfea63c6c3fc6a122aac7a1511c4383a6dc5207a4067"} Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.152843 4876 generic.go:334] "Generic (PLEG): container finished" podID="95f47410-f357-47ce-9f29-f6948f02ed45" containerID="d920151cf960cd9011266a42d377ced5f63cc0564624c236cf81e9406ae2903b" exitCode=0 Mar 13 07:58:46 crc kubenswrapper[4876]: I0313 07:58:46.152904 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-j4z2h" event={"ID":"95f47410-f357-47ce-9f29-f6948f02ed45","Type":"ContainerDied","Data":"d920151cf960cd9011266a42d377ced5f63cc0564624c236cf81e9406ae2903b"} Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.301861 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-654b7b977c-7x6lq"] Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.347785 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66dd85954b-6789x"] Mar 13 07:58:48 crc kubenswrapper[4876]: E0313 07:58:48.348511 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b81d3cbc-3656-4207-adad-46b11ed67d92" containerName="init" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.348536 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b81d3cbc-3656-4207-adad-46b11ed67d92" containerName="init" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.348764 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b81d3cbc-3656-4207-adad-46b11ed67d92" containerName="init" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.349905 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.352868 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.373337 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66dd85954b-6789x"] Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.403720 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5df7599fc5-4djfw"] Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.425813 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-scripts\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.426039 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-config-data\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.426103 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66zhr\" (UniqueName: \"kubernetes.io/projected/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-kube-api-access-66zhr\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.426203 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-combined-ca-bundle\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.426409 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-logs\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.426445 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-secret-key\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.426486 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-tls-certs\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.464583 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-764bff8b56-4l8lg"] Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.467674 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.478994 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-764bff8b56-4l8lg"] Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529094 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-horizon-secret-key\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529148 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70b62196-7476-4c85-a2d7-2299da069989-scripts\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529207 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-logs\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529295 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70b62196-7476-4c85-a2d7-2299da069989-config-data\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529323 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-secret-key\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529382 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-tls-certs\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529426 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-horizon-tls-certs\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529920 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-logs\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.530361 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-scripts\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.529455 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-scripts\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.530697 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcpxl\" (UniqueName: \"kubernetes.io/projected/70b62196-7476-4c85-a2d7-2299da069989-kube-api-access-bcpxl\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.530742 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-config-data\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.530781 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66zhr\" (UniqueName: \"kubernetes.io/projected/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-kube-api-access-66zhr\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.530809 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b62196-7476-4c85-a2d7-2299da069989-logs\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.530848 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-combined-ca-bundle\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.530917 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-combined-ca-bundle\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.532062 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-config-data\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.538547 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-combined-ca-bundle\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.541703 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-tls-certs\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.549415 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-secret-key\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.557100 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66zhr\" (UniqueName: \"kubernetes.io/projected/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-kube-api-access-66zhr\") pod \"horizon-66dd85954b-6789x\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.633685 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-horizon-secret-key\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.633754 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70b62196-7476-4c85-a2d7-2299da069989-scripts\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.633793 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70b62196-7476-4c85-a2d7-2299da069989-config-data\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.633842 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-horizon-tls-certs\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.633890 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcpxl\" (UniqueName: \"kubernetes.io/projected/70b62196-7476-4c85-a2d7-2299da069989-kube-api-access-bcpxl\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.633957 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b62196-7476-4c85-a2d7-2299da069989-logs\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.633988 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-combined-ca-bundle\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.636014 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b62196-7476-4c85-a2d7-2299da069989-logs\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.636076 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70b62196-7476-4c85-a2d7-2299da069989-scripts\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.637407 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70b62196-7476-4c85-a2d7-2299da069989-config-data\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.639477 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-horizon-secret-key\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.642529 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-combined-ca-bundle\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.643697 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/70b62196-7476-4c85-a2d7-2299da069989-horizon-tls-certs\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.661019 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcpxl\" (UniqueName: \"kubernetes.io/projected/70b62196-7476-4c85-a2d7-2299da069989-kube-api-access-bcpxl\") pod \"horizon-764bff8b56-4l8lg\" (UID: \"70b62196-7476-4c85-a2d7-2299da069989\") " pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.715358 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:58:48 crc kubenswrapper[4876]: I0313 07:58:48.814065 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:58:50 crc kubenswrapper[4876]: I0313 07:58:50.326413 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:58:50 crc kubenswrapper[4876]: I0313 07:58:50.391505 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-crr5d"] Mar 13 07:58:50 crc kubenswrapper[4876]: I0313 07:58:50.392038 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="dnsmasq-dns" containerID="cri-o://4fd6fad89a28c78e7cd9a7afe90312ce4fbec880252f8ff385bc6a6bf26aefe0" gracePeriod=10 Mar 13 07:58:51 crc kubenswrapper[4876]: I0313 07:58:51.221472 4876 generic.go:334] "Generic (PLEG): container finished" podID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerID="4fd6fad89a28c78e7cd9a7afe90312ce4fbec880252f8ff385bc6a6bf26aefe0" exitCode=0 Mar 13 07:58:51 crc kubenswrapper[4876]: I0313 07:58:51.221549 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" event={"ID":"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6","Type":"ContainerDied","Data":"4fd6fad89a28c78e7cd9a7afe90312ce4fbec880252f8ff385bc6a6bf26aefe0"} Mar 13 07:58:51 crc kubenswrapper[4876]: I0313 07:58:51.950497 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Mar 13 07:58:54 crc kubenswrapper[4876]: I0313 07:58:54.607693 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:58:54 crc kubenswrapper[4876]: I0313 07:58:54.608090 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:58:56 crc kubenswrapper[4876]: E0313 07:58:56.411643 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 13 07:58:56 crc kubenswrapper[4876]: E0313 07:58:56.411897 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dh58h579h688h658h588h5dbh599hf4h679h575h5d6h56bh694hd7h5fch54fhf9h66fh565h55fh589hd9h567h577h585h5fch57ch696h699h58h56dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lxfxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5df7599fc5-4djfw_openstack(210baf38-0b27-4268-a313-fc13d56f55e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:58:56 crc kubenswrapper[4876]: E0313 07:58:56.421681 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5df7599fc5-4djfw" podUID="210baf38-0b27-4268-a313-fc13d56f55e1" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.515550 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.634623 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-combined-ca-bundle\") pod \"95f47410-f357-47ce-9f29-f6948f02ed45\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.634667 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-scripts\") pod \"95f47410-f357-47ce-9f29-f6948f02ed45\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.634910 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-fernet-keys\") pod \"95f47410-f357-47ce-9f29-f6948f02ed45\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.634957 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-config-data\") pod \"95f47410-f357-47ce-9f29-f6948f02ed45\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.634996 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5v6b\" (UniqueName: \"kubernetes.io/projected/95f47410-f357-47ce-9f29-f6948f02ed45-kube-api-access-n5v6b\") pod \"95f47410-f357-47ce-9f29-f6948f02ed45\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.635039 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-credential-keys\") pod \"95f47410-f357-47ce-9f29-f6948f02ed45\" (UID: \"95f47410-f357-47ce-9f29-f6948f02ed45\") " Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.641994 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "95f47410-f357-47ce-9f29-f6948f02ed45" (UID: "95f47410-f357-47ce-9f29-f6948f02ed45"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.642128 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f47410-f357-47ce-9f29-f6948f02ed45-kube-api-access-n5v6b" (OuterVolumeSpecName: "kube-api-access-n5v6b") pod "95f47410-f357-47ce-9f29-f6948f02ed45" (UID: "95f47410-f357-47ce-9f29-f6948f02ed45"). InnerVolumeSpecName "kube-api-access-n5v6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.642274 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-scripts" (OuterVolumeSpecName: "scripts") pod "95f47410-f357-47ce-9f29-f6948f02ed45" (UID: "95f47410-f357-47ce-9f29-f6948f02ed45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.654415 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "95f47410-f357-47ce-9f29-f6948f02ed45" (UID: "95f47410-f357-47ce-9f29-f6948f02ed45"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.664532 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-config-data" (OuterVolumeSpecName: "config-data") pod "95f47410-f357-47ce-9f29-f6948f02ed45" (UID: "95f47410-f357-47ce-9f29-f6948f02ed45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.665603 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95f47410-f357-47ce-9f29-f6948f02ed45" (UID: "95f47410-f357-47ce-9f29-f6948f02ed45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.738000 4876 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.738106 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.738124 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5v6b\" (UniqueName: \"kubernetes.io/projected/95f47410-f357-47ce-9f29-f6948f02ed45-kube-api-access-n5v6b\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.738136 4876 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.738147 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.738155 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f47410-f357-47ce-9f29-f6948f02ed45-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:56 crc kubenswrapper[4876]: I0313 07:58:56.950668 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.295705 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-j4z2h" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.296596 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-j4z2h" event={"ID":"95f47410-f357-47ce-9f29-f6948f02ed45","Type":"ContainerDied","Data":"27b3ec3de9e26dd4eed47461e95f16621920baf010000cd68ed4a3a9c496e178"} Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.296635 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27b3ec3de9e26dd4eed47461e95f16621920baf010000cd68ed4a3a9c496e178" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.614168 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-j4z2h"] Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.623251 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-j4z2h"] Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.715101 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dtrx5"] Mar 13 07:58:57 crc kubenswrapper[4876]: E0313 07:58:57.717852 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f47410-f357-47ce-9f29-f6948f02ed45" containerName="keystone-bootstrap" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.717888 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f47410-f357-47ce-9f29-f6948f02ed45" containerName="keystone-bootstrap" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.718936 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f47410-f357-47ce-9f29-f6948f02ed45" containerName="keystone-bootstrap" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.720434 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.723889 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.724210 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.726147 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.726600 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-btt8h" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.731965 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.749003 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dtrx5"] Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.872149 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-fernet-keys\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.872301 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9gr9\" (UniqueName: \"kubernetes.io/projected/88446f52-6434-4754-b225-3e79cbccac1b-kube-api-access-n9gr9\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.872397 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-config-data\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.872430 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-credential-keys\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.872454 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-scripts\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.872604 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-combined-ca-bundle\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.974475 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9gr9\" (UniqueName: \"kubernetes.io/projected/88446f52-6434-4754-b225-3e79cbccac1b-kube-api-access-n9gr9\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.974529 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-config-data\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.974553 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-credential-keys\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.974572 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-scripts\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.974634 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-combined-ca-bundle\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.974728 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-fernet-keys\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.981572 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-scripts\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.983428 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-config-data\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.983688 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-credential-keys\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.985184 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-combined-ca-bundle\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.986197 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-fernet-keys\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:57 crc kubenswrapper[4876]: I0313 07:58:57.994207 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9gr9\" (UniqueName: \"kubernetes.io/projected/88446f52-6434-4754-b225-3e79cbccac1b-kube-api-access-n9gr9\") pod \"keystone-bootstrap-dtrx5\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.048722 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:58:58 crc kubenswrapper[4876]: E0313 07:58:58.271449 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Mar 13 07:58:58 crc kubenswrapper[4876]: E0313 07:58:58.271950 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-27g6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-l75wd_openstack(669de0e1-d1a3-4298-bee2-e3713fd7f652): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:58:58 crc kubenswrapper[4876]: E0313 07:58:58.273396 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-l75wd" podUID="669de0e1-d1a3-4298-bee2-e3713fd7f652" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.311347 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5df7599fc5-4djfw" event={"ID":"210baf38-0b27-4268-a313-fc13d56f55e1","Type":"ContainerDied","Data":"7f1b2d62ae4e0c155abced5046679754bc23661e9124c16d9d09f846e123019b"} Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.311410 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f1b2d62ae4e0c155abced5046679754bc23661e9124c16d9d09f846e123019b" Mar 13 07:58:58 crc kubenswrapper[4876]: E0313 07:58:58.313522 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-l75wd" podUID="669de0e1-d1a3-4298-bee2-e3713fd7f652" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.375402 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.483939 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-scripts\") pod \"210baf38-0b27-4268-a313-fc13d56f55e1\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.484177 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/210baf38-0b27-4268-a313-fc13d56f55e1-horizon-secret-key\") pod \"210baf38-0b27-4268-a313-fc13d56f55e1\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.484321 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-config-data\") pod \"210baf38-0b27-4268-a313-fc13d56f55e1\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.484394 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/210baf38-0b27-4268-a313-fc13d56f55e1-logs\") pod \"210baf38-0b27-4268-a313-fc13d56f55e1\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.484422 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxfxt\" (UniqueName: \"kubernetes.io/projected/210baf38-0b27-4268-a313-fc13d56f55e1-kube-api-access-lxfxt\") pod \"210baf38-0b27-4268-a313-fc13d56f55e1\" (UID: \"210baf38-0b27-4268-a313-fc13d56f55e1\") " Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.484501 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-scripts" (OuterVolumeSpecName: "scripts") pod "210baf38-0b27-4268-a313-fc13d56f55e1" (UID: "210baf38-0b27-4268-a313-fc13d56f55e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.484833 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/210baf38-0b27-4268-a313-fc13d56f55e1-logs" (OuterVolumeSpecName: "logs") pod "210baf38-0b27-4268-a313-fc13d56f55e1" (UID: "210baf38-0b27-4268-a313-fc13d56f55e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.485460 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/210baf38-0b27-4268-a313-fc13d56f55e1-logs\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.485498 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.486017 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-config-data" (OuterVolumeSpecName: "config-data") pod "210baf38-0b27-4268-a313-fc13d56f55e1" (UID: "210baf38-0b27-4268-a313-fc13d56f55e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.490384 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210baf38-0b27-4268-a313-fc13d56f55e1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "210baf38-0b27-4268-a313-fc13d56f55e1" (UID: "210baf38-0b27-4268-a313-fc13d56f55e1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.503396 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210baf38-0b27-4268-a313-fc13d56f55e1-kube-api-access-lxfxt" (OuterVolumeSpecName: "kube-api-access-lxfxt") pod "210baf38-0b27-4268-a313-fc13d56f55e1" (UID: "210baf38-0b27-4268-a313-fc13d56f55e1"). InnerVolumeSpecName "kube-api-access-lxfxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.587396 4876 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/210baf38-0b27-4268-a313-fc13d56f55e1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.587446 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/210baf38-0b27-4268-a313-fc13d56f55e1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:58 crc kubenswrapper[4876]: I0313 07:58:58.587462 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxfxt\" (UniqueName: \"kubernetes.io/projected/210baf38-0b27-4268-a313-fc13d56f55e1-kube-api-access-lxfxt\") on node \"crc\" DevicePath \"\"" Mar 13 07:58:59 crc kubenswrapper[4876]: I0313 07:58:59.047786 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f47410-f357-47ce-9f29-f6948f02ed45" path="/var/lib/kubelet/pods/95f47410-f357-47ce-9f29-f6948f02ed45/volumes" Mar 13 07:58:59 crc kubenswrapper[4876]: I0313 07:58:59.317731 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5df7599fc5-4djfw" Mar 13 07:58:59 crc kubenswrapper[4876]: I0313 07:58:59.388021 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5df7599fc5-4djfw"] Mar 13 07:58:59 crc kubenswrapper[4876]: I0313 07:58:59.397616 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5df7599fc5-4djfw"] Mar 13 07:59:00 crc kubenswrapper[4876]: E0313 07:59:00.247873 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 13 07:59:00 crc kubenswrapper[4876]: E0313 07:59:00.248093 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n699h54dh549hc9h597h5d7h586h6bh655h6fh56h5dfh687h5dbh56dh66dh5ch5cfh64fh584h59bhc7h68fh5b5h545h64bh57ch79h698hcchf5hcq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-92grg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-654b7b977c-7x6lq_openstack(bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:59:00 crc kubenswrapper[4876]: E0313 07:59:00.265717 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-654b7b977c-7x6lq" podUID="bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c" Mar 13 07:59:01 crc kubenswrapper[4876]: I0313 07:59:01.050805 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210baf38-0b27-4268-a313-fc13d56f55e1" path="/var/lib/kubelet/pods/210baf38-0b27-4268-a313-fc13d56f55e1/volumes" Mar 13 07:59:03 crc kubenswrapper[4876]: I0313 07:59:03.363142 4876 generic.go:334] "Generic (PLEG): container finished" podID="e0b8464c-522b-44cc-8943-0f0770c2310b" containerID="6c9ccd7f9ccc1773a10fad2983c6fae777e122684bccaaa3d24a91f678f5ce2a" exitCode=0 Mar 13 07:59:03 crc kubenswrapper[4876]: I0313 07:59:03.363625 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hsm7j" event={"ID":"e0b8464c-522b-44cc-8943-0f0770c2310b","Type":"ContainerDied","Data":"6c9ccd7f9ccc1773a10fad2983c6fae777e122684bccaaa3d24a91f678f5ce2a"} Mar 13 07:59:06 crc kubenswrapper[4876]: I0313 07:59:06.950507 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Mar 13 07:59:06 crc kubenswrapper[4876]: I0313 07:59:06.951331 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.130626 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.131139 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tndxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-rzbxc_openstack(86d9e4f3-c1e8-40fe-a6b8-79507f63ca32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.133636 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-rzbxc" podUID="86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.303144 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.320736 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.343639 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.369504 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386269 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-combined-ca-bundle\") pod \"e0b8464c-522b-44cc-8943-0f0770c2310b\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386340 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-svc\") pod \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386409 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386452 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-combined-ca-bundle\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386508 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nrb7\" (UniqueName: \"kubernetes.io/projected/42968848-d1aa-4140-9a19-0218cdbc98d8-kube-api-access-5nrb7\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386550 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb\") pod \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386602 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-internal-tls-certs\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386632 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-nb\") pod \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386678 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7zpq\" (UniqueName: \"kubernetes.io/projected/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-kube-api-access-z7zpq\") pod \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386698 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7scb8\" (UniqueName: \"kubernetes.io/projected/e0b8464c-522b-44cc-8943-0f0770c2310b-kube-api-access-7scb8\") pod \"e0b8464c-522b-44cc-8943-0f0770c2310b\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386730 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-config\") pod \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386778 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-swift-storage-0\") pod \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386801 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-scripts\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386825 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-logs\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386873 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-config-data\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.386965 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-httpd-run\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.387002 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-config\") pod \"e0b8464c-522b-44cc-8943-0f0770c2310b\" (UID: \"e0b8464c-522b-44cc-8943-0f0770c2310b\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.387559 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.405637 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0b8464c-522b-44cc-8943-0f0770c2310b-kube-api-access-7scb8" (OuterVolumeSpecName: "kube-api-access-7scb8") pod "e0b8464c-522b-44cc-8943-0f0770c2310b" (UID: "e0b8464c-522b-44cc-8943-0f0770c2310b"). InnerVolumeSpecName "kube-api-access-7scb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.405692 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-kube-api-access-z7zpq" (OuterVolumeSpecName: "kube-api-access-z7zpq") pod "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" (UID: "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6"). InnerVolumeSpecName "kube-api-access-z7zpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.405973 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-logs" (OuterVolumeSpecName: "logs") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.406039 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42968848-d1aa-4140-9a19-0218cdbc98d8-kube-api-access-5nrb7" (OuterVolumeSpecName: "kube-api-access-5nrb7") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "kube-api-access-5nrb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.406766 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.409545 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.434065 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" event={"ID":"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6","Type":"ContainerDied","Data":"f072db7b78651a7b2d7a09b36e14ce6bc1ab7abe9f4259888ba0a2a95036b139"} Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.434128 4876 scope.go:117] "RemoveContainer" containerID="4fd6fad89a28c78e7cd9a7afe90312ce4fbec880252f8ff385bc6a6bf26aefe0" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.434257 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.446488 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-scripts" (OuterVolumeSpecName: "scripts") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.452840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.455729 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34","Type":"ContainerDied","Data":"407adf5d98657b8252d8abcf1f1b56219361c33bb87e3a4829c978690e98b234"} Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.455835 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.463904 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"42968848-d1aa-4140-9a19-0218cdbc98d8","Type":"ContainerDied","Data":"0c6749d256b69aa33d06ffeb3ec1ee909b5a65ded73db1cd690e68222b2c69c0"} Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.463994 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.467107 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-654b7b977c-7x6lq" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.467196 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b7b977c-7x6lq" event={"ID":"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c","Type":"ContainerDied","Data":"69069459a42e0807683906e45d5ffff6fa3ff077a928cb5beb561d8c276ae755"} Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.473647 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hsm7j" event={"ID":"e0b8464c-522b-44cc-8943-0f0770c2310b","Type":"ContainerDied","Data":"a72baf22ee60aadf4f309a3915fc3a280afe39d744b04ca045c6f063d397862c"} Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.473737 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a72baf22ee60aadf4f309a3915fc3a280afe39d744b04ca045c6f063d397862c" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.473920 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hsm7j" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.475338 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-rzbxc" podUID="86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.489024 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-logs\") pod \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.489461 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-scripts\") pod \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.489960 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92grg\" (UniqueName: \"kubernetes.io/projected/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-kube-api-access-92grg\") pod \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.490083 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-combined-ca-bundle\") pod \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.490283 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t24vb\" (UniqueName: \"kubernetes.io/projected/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-kube-api-access-t24vb\") pod \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.490494 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-public-tls-certs\") pod \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.490578 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-horizon-secret-key\") pod \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.490724 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.490840 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-config-data\") pod \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.490946 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-scripts\") pod \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.491099 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-config-data\") pod \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.491215 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-httpd-run\") pod \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\" (UID: \"9b451bf9-4e5a-4521-a49c-2f2e7f72eb34\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.491348 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-logs\") pod \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\" (UID: \"bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c\") " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.491861 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.491949 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.492040 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.492110 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nrb7\" (UniqueName: \"kubernetes.io/projected/42968848-d1aa-4140-9a19-0218cdbc98d8-kube-api-access-5nrb7\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.492170 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7zpq\" (UniqueName: \"kubernetes.io/projected/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-kube-api-access-z7zpq\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.492291 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7scb8\" (UniqueName: \"kubernetes.io/projected/e0b8464c-522b-44cc-8943-0f0770c2310b-kube-api-access-7scb8\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.492401 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.492474 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42968848-d1aa-4140-9a19-0218cdbc98d8-logs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.490443 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0b8464c-522b-44cc-8943-0f0770c2310b" (UID: "e0b8464c-522b-44cc-8943-0f0770c2310b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.494829 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" (UID: "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.494858 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-logs" (OuterVolumeSpecName: "logs") pod "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" (UID: "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.495193 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-logs" (OuterVolumeSpecName: "logs") pod "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c" (UID: "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.495486 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c" (UID: "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.496754 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-scripts" (OuterVolumeSpecName: "scripts") pod "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c" (UID: "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.502287 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-config" (OuterVolumeSpecName: "config") pod "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" (UID: "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.502617 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-config-data" (OuterVolumeSpecName: "config-data") pod "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c" (UID: "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.502653 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-config" (OuterVolumeSpecName: "config") pod "e0b8464c-522b-44cc-8943-0f0770c2310b" (UID: "e0b8464c-522b-44cc-8943-0f0770c2310b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.508843 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-kube-api-access-t24vb" (OuterVolumeSpecName: "kube-api-access-t24vb") pod "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" (UID: "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34"). InnerVolumeSpecName "kube-api-access-t24vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.514425 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-kube-api-access-92grg" (OuterVolumeSpecName: "kube-api-access-92grg") pod "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c" (UID: "bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c"). InnerVolumeSpecName "kube-api-access-92grg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.514932 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-scripts" (OuterVolumeSpecName: "scripts") pod "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" (UID: "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.527702 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" (UID: "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.531831 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" (UID: "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.560706 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" (UID: "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.568928 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.572080 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" (UID: "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.578412 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb podName:e6a54139-e072-46e7-b5e5-a1f0e52a0ca6 nodeName:}" failed. No retries permitted until 2026-03-13 07:59:09.078381131 +0000 UTC m=+1208.749160313 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-sb" (UniqueName: "kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb") pod "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" (UID: "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6") : error deleting /var/lib/kubelet/pods/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6/volume-subpaths: remove /var/lib/kubelet/pods/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6/volume-subpaths: no such file or directory Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.579028 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" (UID: "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.592882 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.593670 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-internal-tls-certs\") pod \"42968848-d1aa-4140-9a19-0218cdbc98d8\" (UID: \"42968848-d1aa-4140-9a19-0218cdbc98d8\") " Mar 13 07:59:08 crc kubenswrapper[4876]: W0313 07:59:08.593819 4876 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/42968848-d1aa-4140-9a19-0218cdbc98d8/volumes/kubernetes.io~secret/internal-tls-certs Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.593842 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.594583 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.594660 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.594719 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.594775 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92grg\" (UniqueName: \"kubernetes.io/projected/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-kube-api-access-92grg\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.594842 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.594900 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t24vb\" (UniqueName: \"kubernetes.io/projected/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-kube-api-access-t24vb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.594954 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595007 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595057 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595113 4876 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595167 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595249 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595316 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595368 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595420 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595475 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c-logs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595528 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595580 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0b8464c-522b-44cc-8943-0f0770c2310b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.595639 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-logs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.601401 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-config-data" (OuterVolumeSpecName: "config-data") pod "42968848-d1aa-4140-9a19-0218cdbc98d8" (UID: "42968848-d1aa-4140-9a19-0218cdbc98d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.616332 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" (UID: "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.618161 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-config-data" (OuterVolumeSpecName: "config-data") pod "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" (UID: "9b451bf9-4e5a-4521-a49c-2f2e7f72eb34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.619144 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.697483 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42968848-d1aa-4140-9a19-0218cdbc98d8-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.697519 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.697530 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.697542 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.811560 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.835589 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.853986 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.868932 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.878461 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.878946 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="dnsmasq-dns" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.878969 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="dnsmasq-dns" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.878979 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerName="glance-httpd" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.878987 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerName="glance-httpd" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.879003 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerName="glance-log" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879010 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerName="glance-log" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.879031 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerName="glance-log" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879037 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerName="glance-log" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.879050 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="init" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879059 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="init" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.879074 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerName="glance-httpd" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879080 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerName="glance-httpd" Mar 13 07:59:08 crc kubenswrapper[4876]: E0313 07:59:08.879092 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0b8464c-522b-44cc-8943-0f0770c2310b" containerName="neutron-db-sync" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879098 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0b8464c-522b-44cc-8943-0f0770c2310b" containerName="neutron-db-sync" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879273 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerName="glance-httpd" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879289 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerName="glance-log" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879298 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" containerName="glance-log" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879310 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0b8464c-522b-44cc-8943-0f0770c2310b" containerName="neutron-db-sync" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879318 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="dnsmasq-dns" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.879326 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" containerName="glance-httpd" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.880496 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.885267 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.885354 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.885913 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.888343 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.888660 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xxnmv" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.890491 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.894387 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.894647 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.903184 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.950137 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.977484 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-654b7b977c-7x6lq"] Mar 13 07:59:08 crc kubenswrapper[4876]: I0313 07:59:08.982942 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-654b7b977c-7x6lq"] Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015483 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015538 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk264\" (UniqueName: \"kubernetes.io/projected/17347053-1f92-46e8-a1e2-be471cf19516-kube-api-access-dk264\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015579 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015613 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015629 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-config-data\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015752 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015808 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015878 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015899 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.015963 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.016032 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-logs\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.016050 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.016077 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.016181 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x22nj\" (UniqueName: \"kubernetes.io/projected/b3664dea-7307-496d-aff3-588911d155bb-kube-api-access-x22nj\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.016214 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.016284 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-scripts\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.048630 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42968848-d1aa-4140-9a19-0218cdbc98d8" path="/var/lib/kubelet/pods/42968848-d1aa-4140-9a19-0218cdbc98d8/volumes" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.049893 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b451bf9-4e5a-4521-a49c-2f2e7f72eb34" path="/var/lib/kubelet/pods/9b451bf9-4e5a-4521-a49c-2f2e7f72eb34/volumes" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.050790 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c" path="/var/lib/kubelet/pods/bf1e74fd-3e8e-405c-8fc3-3f095b7f8c7c/volumes" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118032 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb\") pod \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\" (UID: \"e6a54139-e072-46e7-b5e5-a1f0e52a0ca6\") " Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118528 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118602 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118605 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" (UID: "e6a54139-e072-46e7-b5e5-a1f0e52a0ca6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118625 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118738 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118822 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-logs\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118845 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118883 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.118985 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x22nj\" (UniqueName: \"kubernetes.io/projected/b3664dea-7307-496d-aff3-588911d155bb-kube-api-access-x22nj\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.119017 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.119045 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.119073 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-scripts\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.119077 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.119200 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.119843 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-logs\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.120289 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.120490 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.120552 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk264\" (UniqueName: \"kubernetes.io/projected/17347053-1f92-46e8-a1e2-be471cf19516-kube-api-access-dk264\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.120676 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.120789 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.120814 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-config-data\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.120846 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.120951 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.122852 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.123466 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.123556 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.125312 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.132387 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.137307 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.139386 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-scripts\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.141544 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.142893 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x22nj\" (UniqueName: \"kubernetes.io/projected/b3664dea-7307-496d-aff3-588911d155bb-kube-api-access-x22nj\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.143779 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-config-data\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.145113 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk264\" (UniqueName: \"kubernetes.io/projected/17347053-1f92-46e8-a1e2-be471cf19516-kube-api-access-dk264\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.169385 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.171786 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.200576 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.215838 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.402031 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-crr5d"] Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.411675 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-crr5d"] Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.589505 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-cv8h6"] Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.591610 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.600152 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-cv8h6"] Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.609278 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5cb9cf56c6-k7vkg"] Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.610982 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.615148 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.615389 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.615593 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wr4mz" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.615781 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.619681 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5cb9cf56c6-k7vkg"] Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.644497 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.644553 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-config\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.644580 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-svc\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.644637 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.644695 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fxjd\" (UniqueName: \"kubernetes.io/projected/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-kube-api-access-8fxjd\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.644820 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746631 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746726 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746766 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-config\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746797 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89mmv\" (UniqueName: \"kubernetes.io/projected/9fd26454-5f72-4e52-a3e7-87270228d46d-kube-api-access-89mmv\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746830 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-svc\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746859 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-ovndb-tls-certs\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746912 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746944 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-config\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.746990 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fxjd\" (UniqueName: \"kubernetes.io/projected/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-kube-api-access-8fxjd\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.747029 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-combined-ca-bundle\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.747065 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-httpd-config\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.748070 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.748752 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.749409 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-config\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.749611 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.749997 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-svc\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.767502 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fxjd\" (UniqueName: \"kubernetes.io/projected/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-kube-api-access-8fxjd\") pod \"dnsmasq-dns-55f844cf75-cv8h6\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.849414 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-config\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.849508 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-combined-ca-bundle\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.849540 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-httpd-config\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.849615 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89mmv\" (UniqueName: \"kubernetes.io/projected/9fd26454-5f72-4e52-a3e7-87270228d46d-kube-api-access-89mmv\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.849640 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-ovndb-tls-certs\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.853613 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-httpd-config\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.853811 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-combined-ca-bundle\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.855975 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-ovndb-tls-certs\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.863963 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-config\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.868674 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89mmv\" (UniqueName: \"kubernetes.io/projected/9fd26454-5f72-4e52-a3e7-87270228d46d-kube-api-access-89mmv\") pod \"neutron-5cb9cf56c6-k7vkg\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.974416 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:09 crc kubenswrapper[4876]: I0313 07:59:09.986751 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:10 crc kubenswrapper[4876]: I0313 07:59:10.296566 4876 scope.go:117] "RemoveContainer" containerID="45fc2002c0203cc2e54785692fc0ba68400f3f50db1ce89d4c4fade21ed6ac0e" Mar 13 07:59:10 crc kubenswrapper[4876]: E0313 07:59:10.393744 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 13 07:59:10 crc kubenswrapper[4876]: E0313 07:59:10.394260 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sz79q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-ltwdh_openstack(5d42486d-85b1-4384-91fc-10552a2c404b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 07:59:10 crc kubenswrapper[4876]: E0313 07:59:10.395484 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-ltwdh" podUID="5d42486d-85b1-4384-91fc-10552a2c404b" Mar 13 07:59:10 crc kubenswrapper[4876]: E0313 07:59:10.522682 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-ltwdh" podUID="5d42486d-85b1-4384-91fc-10552a2c404b" Mar 13 07:59:10 crc kubenswrapper[4876]: I0313 07:59:10.547498 4876 scope.go:117] "RemoveContainer" containerID="89d2bc28ca15c6767b57fe40afe46636199325610ce16c3bbab01bd019fdcc49" Mar 13 07:59:10 crc kubenswrapper[4876]: I0313 07:59:10.755618 4876 scope.go:117] "RemoveContainer" containerID="9ba0d1199b1516989333cfea63c6c3fc6a122aac7a1511c4383a6dc5207a4067" Mar 13 07:59:10 crc kubenswrapper[4876]: I0313 07:59:10.788443 4876 scope.go:117] "RemoveContainer" containerID="3623ef20ce0509f3a529da1f907a34545f69dea1c16c361e96d6ee4a558c49c3" Mar 13 07:59:10 crc kubenswrapper[4876]: I0313 07:59:10.806802 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66dd85954b-6789x"] Mar 13 07:59:10 crc kubenswrapper[4876]: I0313 07:59:10.849793 4876 scope.go:117] "RemoveContainer" containerID="43eb9beea83cea81bfb0030fd545e7ecb5558c5a0ceb54835ff5e8cdaa8bfae4" Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.005862 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dtrx5"] Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.017263 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.029138 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-764bff8b56-4l8lg"] Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.059643 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" path="/var/lib/kubelet/pods/e6a54139-e072-46e7-b5e5-a1f0e52a0ca6/volumes" Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.218546 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-cv8h6"] Mar 13 07:59:11 crc kubenswrapper[4876]: W0313 07:59:11.317737 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fd26454_5f72_4e52_a3e7_87270228d46d.slice/crio-4b668cfe732b0d00c37d1335159394536244fe6a51656937b48dc42867d7e278 WatchSource:0}: Error finding container 4b668cfe732b0d00c37d1335159394536244fe6a51656937b48dc42867d7e278: Status 404 returned error can't find the container with id 4b668cfe732b0d00c37d1335159394536244fe6a51656937b48dc42867d7e278 Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.326682 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5cb9cf56c6-k7vkg"] Mar 13 07:59:11 crc kubenswrapper[4876]: W0313 07:59:11.454417 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3664dea_7307_496d_aff3_588911d155bb.slice/crio-67467ecebda852dfc6fd0eaf843f612964966aa301a9bbc020cad4c65d98e79c WatchSource:0}: Error finding container 67467ecebda852dfc6fd0eaf843f612964966aa301a9bbc020cad4c65d98e79c: Status 404 returned error can't find the container with id 67467ecebda852dfc6fd0eaf843f612964966aa301a9bbc020cad4c65d98e79c Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.476686 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.563366 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dtrx5" event={"ID":"88446f52-6434-4754-b225-3e79cbccac1b","Type":"ContainerStarted","Data":"396ac8d43845968d60c9c9a325f780eaa34e76d7d1cb79e591f45a4d0cb927d7"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.563438 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dtrx5" event={"ID":"88446f52-6434-4754-b225-3e79cbccac1b","Type":"ContainerStarted","Data":"66080dbd64209128231a8ea881031c7e752baa2ac26805332e1e84b53c583a7c"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.595781 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dtrx5" podStartSLOduration=14.595756641 podStartE2EDuration="14.595756641s" podCreationTimestamp="2026-03-13 07:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:11.593757503 +0000 UTC m=+1211.264536485" watchObservedRunningTime="2026-03-13 07:59:11.595756641 +0000 UTC m=+1211.266535623" Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.644746 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59996cc657-lxnh9" event={"ID":"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a","Type":"ContainerStarted","Data":"0bebe7f9d85929a662bbc5ed96fc28b1d498412a36ff797ea0962b227bd9884d"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.644826 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59996cc657-lxnh9" event={"ID":"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a","Type":"ContainerStarted","Data":"289009ab13035d7e67f3fef975563573dcbd7b63473eda841713899882a68394"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.645079 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-59996cc657-lxnh9" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerName="horizon-log" containerID="cri-o://289009ab13035d7e67f3fef975563573dcbd7b63473eda841713899882a68394" gracePeriod=30 Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.648773 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-59996cc657-lxnh9" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerName="horizon" containerID="cri-o://0bebe7f9d85929a662bbc5ed96fc28b1d498412a36ff797ea0962b227bd9884d" gracePeriod=30 Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.692681 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-764bff8b56-4l8lg" event={"ID":"70b62196-7476-4c85-a2d7-2299da069989","Type":"ContainerStarted","Data":"7dd937256ac090bd1be94054a0c11333b432444f5c05ef29e25344dbb7101ff2"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.692793 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-764bff8b56-4l8lg" event={"ID":"70b62196-7476-4c85-a2d7-2299da069989","Type":"ContainerStarted","Data":"fdb80b7eceaa6249bbb240864b0541b124054a9d91fc761318ebbaebc0a49cf6"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.712121 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-59996cc657-lxnh9" podStartSLOduration=5.667256311 podStartE2EDuration="32.712103487s" podCreationTimestamp="2026-03-13 07:58:39 +0000 UTC" firstStartedPulling="2026-03-13 07:58:41.130619746 +0000 UTC m=+1180.801398728" lastFinishedPulling="2026-03-13 07:59:08.175466902 +0000 UTC m=+1207.846245904" observedRunningTime="2026-03-13 07:59:11.694840905 +0000 UTC m=+1211.365619897" watchObservedRunningTime="2026-03-13 07:59:11.712103487 +0000 UTC m=+1211.382882469" Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.724629 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cb9cf56c6-k7vkg" event={"ID":"9fd26454-5f72-4e52-a3e7-87270228d46d","Type":"ContainerStarted","Data":"4b668cfe732b0d00c37d1335159394536244fe6a51656937b48dc42867d7e278"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.772600 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66dd85954b-6789x" event={"ID":"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163","Type":"ContainerStarted","Data":"b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.772676 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66dd85954b-6789x" event={"ID":"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163","Type":"ContainerStarted","Data":"fc61cb1cc6bb2cd2db65514e555a97acf5d04e079d61d9a8f67f8b321e75e6fb"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.775121 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" event={"ID":"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1","Type":"ContainerStarted","Data":"32a43100d669dd65512a75a5eb896be277937b64ac0913ce1613a22bce1c84a9"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.785717 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerStarted","Data":"2e99ade3cdbc5460384ad6c59b414c3d139f3f8b1523d2abf7cd47152566ea2a"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.797670 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3664dea-7307-496d-aff3-588911d155bb","Type":"ContainerStarted","Data":"67467ecebda852dfc6fd0eaf843f612964966aa301a9bbc020cad4c65d98e79c"} Mar 13 07:59:11 crc kubenswrapper[4876]: I0313 07:59:11.955163 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-crr5d" podUID="e6a54139-e072-46e7-b5e5-a1f0e52a0ca6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.178424 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.653323 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-78dc766d77-2lq9j"] Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.656384 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.671952 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.685333 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.710047 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78dc766d77-2lq9j"] Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.768931 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-ovndb-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.769004 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-994tn\" (UniqueName: \"kubernetes.io/projected/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-kube-api-access-994tn\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.769076 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-config\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.769099 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-internal-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.769124 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-httpd-config\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.769142 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-public-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.769173 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-combined-ca-bundle\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.833557 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66dd85954b-6789x" event={"ID":"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163","Type":"ContainerStarted","Data":"645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74"} Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.846964 4876 generic.go:334] "Generic (PLEG): container finished" podID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" containerID="f38dea91a9d754ab4d9689a4afc91a181cdc068b29cd15f1a6e47c49200cb625" exitCode=0 Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.847034 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" event={"ID":"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1","Type":"ContainerDied","Data":"f38dea91a9d754ab4d9689a4afc91a181cdc068b29cd15f1a6e47c49200cb625"} Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.856972 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17347053-1f92-46e8-a1e2-be471cf19516","Type":"ContainerStarted","Data":"f9e9233c31d37107a2d6b9e2c76b0d5bfcbfc88ed0bac35e270d482eb23f6d8c"} Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.859119 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cb9cf56c6-k7vkg" event={"ID":"9fd26454-5f72-4e52-a3e7-87270228d46d","Type":"ContainerStarted","Data":"50e23ad80caa9b0a975b0d31c0f0ca3c2d7d59e5b837f13c9b725b68dbce8a4e"} Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.859154 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cb9cf56c6-k7vkg" event={"ID":"9fd26454-5f72-4e52-a3e7-87270228d46d","Type":"ContainerStarted","Data":"78a12f079c86c03a1434cab042a6e40ad549af96a3e8a139ac0f74fb201bedce"} Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.859408 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.871125 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-ovndb-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.871201 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-994tn\" (UniqueName: \"kubernetes.io/projected/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-kube-api-access-994tn\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.871301 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-config\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.871332 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-internal-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.871368 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-httpd-config\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.871392 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-public-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.871440 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-combined-ca-bundle\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.872740 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-764bff8b56-4l8lg" event={"ID":"70b62196-7476-4c85-a2d7-2299da069989","Type":"ContainerStarted","Data":"4b916ca576eb7da1149dd10768e40029a80d406976acef19e24daf3893e5df57"} Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.873647 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66dd85954b-6789x" podStartSLOduration=24.873618641 podStartE2EDuration="24.873618641s" podCreationTimestamp="2026-03-13 07:58:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:12.863208105 +0000 UTC m=+1212.533987087" watchObservedRunningTime="2026-03-13 07:59:12.873618641 +0000 UTC m=+1212.544397623" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.890894 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-public-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.892086 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-httpd-config\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.894796 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-combined-ca-bundle\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.895047 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-internal-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.903152 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-config\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.903761 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-ovndb-tls-certs\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.930819 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-994tn\" (UniqueName: \"kubernetes.io/projected/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-kube-api-access-994tn\") pod \"neutron-78dc766d77-2lq9j\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.965519 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5cb9cf56c6-k7vkg" podStartSLOduration=3.965493612 podStartE2EDuration="3.965493612s" podCreationTimestamp="2026-03-13 07:59:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:12.919042342 +0000 UTC m=+1212.589821324" watchObservedRunningTime="2026-03-13 07:59:12.965493612 +0000 UTC m=+1212.636272594" Mar 13 07:59:12 crc kubenswrapper[4876]: I0313 07:59:12.982664 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-764bff8b56-4l8lg" podStartSLOduration=24.982637659 podStartE2EDuration="24.982637659s" podCreationTimestamp="2026-03-13 07:58:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:12.959310127 +0000 UTC m=+1212.630089119" watchObservedRunningTime="2026-03-13 07:59:12.982637659 +0000 UTC m=+1212.653416641" Mar 13 07:59:13 crc kubenswrapper[4876]: I0313 07:59:13.078181 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:13 crc kubenswrapper[4876]: I0313 07:59:13.924985 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" event={"ID":"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1","Type":"ContainerStarted","Data":"374ddce7f5025c68c1d9607cb69dd15de70e48ec3b5796bf7fa40ff95677c903"} Mar 13 07:59:13 crc kubenswrapper[4876]: I0313 07:59:13.925678 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:13 crc kubenswrapper[4876]: I0313 07:59:13.958058 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17347053-1f92-46e8-a1e2-be471cf19516","Type":"ContainerStarted","Data":"67d5719f159209f2dad8684db9e7b94b6e9dc28cd3cc9f5d8b5c4ddf4962f041"} Mar 13 07:59:13 crc kubenswrapper[4876]: I0313 07:59:13.977751 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3664dea-7307-496d-aff3-588911d155bb","Type":"ContainerStarted","Data":"8862cfa4ed45d9337b0f1fd8dc684f0790bbcc8d758836f140c02da1ae2631ea"} Mar 13 07:59:13 crc kubenswrapper[4876]: I0313 07:59:13.990138 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" podStartSLOduration=4.990104397 podStartE2EDuration="4.990104397s" podCreationTimestamp="2026-03-13 07:59:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:13.958474707 +0000 UTC m=+1213.629253709" watchObservedRunningTime="2026-03-13 07:59:13.990104397 +0000 UTC m=+1213.660883389" Mar 13 07:59:13 crc kubenswrapper[4876]: I0313 07:59:13.996803 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-l75wd" event={"ID":"669de0e1-d1a3-4298-bee2-e3713fd7f652","Type":"ContainerStarted","Data":"849536096db85d5fccb4972e93f7f2407fb4aea0e532aaed84997018aa2c3d03"} Mar 13 07:59:14 crc kubenswrapper[4876]: I0313 07:59:14.154613 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-l75wd" podStartSLOduration=3.771206554 podStartE2EDuration="35.15459093s" podCreationTimestamp="2026-03-13 07:58:39 +0000 UTC" firstStartedPulling="2026-03-13 07:58:41.111909915 +0000 UTC m=+1180.782688907" lastFinishedPulling="2026-03-13 07:59:12.495294301 +0000 UTC m=+1212.166073283" observedRunningTime="2026-03-13 07:59:14.038006438 +0000 UTC m=+1213.708785420" watchObservedRunningTime="2026-03-13 07:59:14.15459093 +0000 UTC m=+1213.825369912" Mar 13 07:59:14 crc kubenswrapper[4876]: I0313 07:59:14.159346 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-78dc766d77-2lq9j"] Mar 13 07:59:15 crc kubenswrapper[4876]: I0313 07:59:15.009611 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3664dea-7307-496d-aff3-588911d155bb","Type":"ContainerStarted","Data":"ac6988d9ae894d1f9bb1fe44270e4e0f5f2765ea43b1cc4650101b6c16699110"} Mar 13 07:59:15 crc kubenswrapper[4876]: I0313 07:59:15.031108 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.031086786 podStartE2EDuration="7.031086786s" podCreationTimestamp="2026-03-13 07:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:15.029069099 +0000 UTC m=+1214.699848081" watchObservedRunningTime="2026-03-13 07:59:15.031086786 +0000 UTC m=+1214.701865768" Mar 13 07:59:15 crc kubenswrapper[4876]: W0313 07:59:15.219473 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod130c3a6d_4650_4cb3_a46b_9a08eb9108e3.slice/crio-4da2c3a76a1815831dbf68cac85263b6ac0549a3f22ddc8d30e4de7d3dd63210 WatchSource:0}: Error finding container 4da2c3a76a1815831dbf68cac85263b6ac0549a3f22ddc8d30e4de7d3dd63210: Status 404 returned error can't find the container with id 4da2c3a76a1815831dbf68cac85263b6ac0549a3f22ddc8d30e4de7d3dd63210 Mar 13 07:59:15 crc kubenswrapper[4876]: I0313 07:59:15.321155 4876 scope.go:117] "RemoveContainer" containerID="1bd84d48a8e6bbb46eff26ac0a20237dbee3e8d502a155a3800850ade3a54e24" Mar 13 07:59:16 crc kubenswrapper[4876]: I0313 07:59:16.054282 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerStarted","Data":"d21cc25bd945c08dee67e81d7642e373e6e49cc67322be5d3e60e5c28f6d5bf5"} Mar 13 07:59:16 crc kubenswrapper[4876]: I0313 07:59:16.059921 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc766d77-2lq9j" event={"ID":"130c3a6d-4650-4cb3-a46b-9a08eb9108e3","Type":"ContainerStarted","Data":"80661cfa70c99dd7b3d101dbcd3d9f87ed202b141eb93e53b24e735b71649c7c"} Mar 13 07:59:16 crc kubenswrapper[4876]: I0313 07:59:16.059961 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc766d77-2lq9j" event={"ID":"130c3a6d-4650-4cb3-a46b-9a08eb9108e3","Type":"ContainerStarted","Data":"4da2c3a76a1815831dbf68cac85263b6ac0549a3f22ddc8d30e4de7d3dd63210"} Mar 13 07:59:17 crc kubenswrapper[4876]: I0313 07:59:17.081440 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17347053-1f92-46e8-a1e2-be471cf19516","Type":"ContainerStarted","Data":"9d9755e38fd10bcc26796d4ab4445627fb53701c36a6fd0b6a7e471def88f7e9"} Mar 13 07:59:17 crc kubenswrapper[4876]: I0313 07:59:17.087027 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc766d77-2lq9j" event={"ID":"130c3a6d-4650-4cb3-a46b-9a08eb9108e3","Type":"ContainerStarted","Data":"a82353ca03963734c9e8d399b80dc22325e344e78ee20af877d27c28fbfe3390"} Mar 13 07:59:17 crc kubenswrapper[4876]: I0313 07:59:17.088157 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:17 crc kubenswrapper[4876]: I0313 07:59:17.113219 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.11319717 podStartE2EDuration="9.11319717s" podCreationTimestamp="2026-03-13 07:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:17.105759348 +0000 UTC m=+1216.776538330" watchObservedRunningTime="2026-03-13 07:59:17.11319717 +0000 UTC m=+1216.783976142" Mar 13 07:59:17 crc kubenswrapper[4876]: I0313 07:59:17.139065 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-78dc766d77-2lq9j" podStartSLOduration=5.139040844 podStartE2EDuration="5.139040844s" podCreationTimestamp="2026-03-13 07:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:17.133909218 +0000 UTC m=+1216.804688200" watchObservedRunningTime="2026-03-13 07:59:17.139040844 +0000 UTC m=+1216.809819826" Mar 13 07:59:18 crc kubenswrapper[4876]: I0313 07:59:18.123091 4876 generic.go:334] "Generic (PLEG): container finished" podID="88446f52-6434-4754-b225-3e79cbccac1b" containerID="396ac8d43845968d60c9c9a325f780eaa34e76d7d1cb79e591f45a4d0cb927d7" exitCode=0 Mar 13 07:59:18 crc kubenswrapper[4876]: I0313 07:59:18.123276 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dtrx5" event={"ID":"88446f52-6434-4754-b225-3e79cbccac1b","Type":"ContainerDied","Data":"396ac8d43845968d60c9c9a325f780eaa34e76d7d1cb79e591f45a4d0cb927d7"} Mar 13 07:59:18 crc kubenswrapper[4876]: I0313 07:59:18.129827 4876 generic.go:334] "Generic (PLEG): container finished" podID="669de0e1-d1a3-4298-bee2-e3713fd7f652" containerID="849536096db85d5fccb4972e93f7f2407fb4aea0e532aaed84997018aa2c3d03" exitCode=0 Mar 13 07:59:18 crc kubenswrapper[4876]: I0313 07:59:18.129897 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-l75wd" event={"ID":"669de0e1-d1a3-4298-bee2-e3713fd7f652","Type":"ContainerDied","Data":"849536096db85d5fccb4972e93f7f2407fb4aea0e532aaed84997018aa2c3d03"} Mar 13 07:59:18 crc kubenswrapper[4876]: I0313 07:59:18.718348 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:59:18 crc kubenswrapper[4876]: I0313 07:59:18.719425 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:59:18 crc kubenswrapper[4876]: I0313 07:59:18.816650 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:59:18 crc kubenswrapper[4876]: I0313 07:59:18.817932 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.201164 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.201628 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.216165 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.216538 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.296818 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.310974 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.311123 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.317885 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.970019 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:59:19 crc kubenswrapper[4876]: I0313 07:59:19.976387 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:20 crc kubenswrapper[4876]: I0313 07:59:20.053193 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5v9sw"] Mar 13 07:59:20 crc kubenswrapper[4876]: I0313 07:59:20.053510 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerName="dnsmasq-dns" containerID="cri-o://36fff9f513672061d505049d00b22d172bdafe59341929dfa7826153aec5016e" gracePeriod=10 Mar 13 07:59:20 crc kubenswrapper[4876]: I0313 07:59:20.153167 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:20 crc kubenswrapper[4876]: I0313 07:59:20.153426 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 07:59:20 crc kubenswrapper[4876]: I0313 07:59:20.153462 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:20 crc kubenswrapper[4876]: I0313 07:59:20.153475 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 07:59:20 crc kubenswrapper[4876]: I0313 07:59:20.327548 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: connect: connection refused" Mar 13 07:59:21 crc kubenswrapper[4876]: I0313 07:59:21.172959 4876 generic.go:334] "Generic (PLEG): container finished" podID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerID="36fff9f513672061d505049d00b22d172bdafe59341929dfa7826153aec5016e" exitCode=0 Mar 13 07:59:21 crc kubenswrapper[4876]: I0313 07:59:21.173053 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" event={"ID":"2b6f7e51-08c6-4e1a-b099-31ffc41866ea","Type":"ContainerDied","Data":"36fff9f513672061d505049d00b22d172bdafe59341929dfa7826153aec5016e"} Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.194473 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.796653 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.832994 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9gr9\" (UniqueName: \"kubernetes.io/projected/88446f52-6434-4754-b225-3e79cbccac1b-kube-api-access-n9gr9\") pod \"88446f52-6434-4754-b225-3e79cbccac1b\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.833171 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-config-data\") pod \"88446f52-6434-4754-b225-3e79cbccac1b\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.833259 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-credential-keys\") pod \"88446f52-6434-4754-b225-3e79cbccac1b\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.833291 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-scripts\") pod \"88446f52-6434-4754-b225-3e79cbccac1b\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.833343 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-combined-ca-bundle\") pod \"88446f52-6434-4754-b225-3e79cbccac1b\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.833501 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-fernet-keys\") pod \"88446f52-6434-4754-b225-3e79cbccac1b\" (UID: \"88446f52-6434-4754-b225-3e79cbccac1b\") " Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.849631 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88446f52-6434-4754-b225-3e79cbccac1b-kube-api-access-n9gr9" (OuterVolumeSpecName: "kube-api-access-n9gr9") pod "88446f52-6434-4754-b225-3e79cbccac1b" (UID: "88446f52-6434-4754-b225-3e79cbccac1b"). InnerVolumeSpecName "kube-api-access-n9gr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.866520 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-scripts" (OuterVolumeSpecName: "scripts") pod "88446f52-6434-4754-b225-3e79cbccac1b" (UID: "88446f52-6434-4754-b225-3e79cbccac1b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.881478 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "88446f52-6434-4754-b225-3e79cbccac1b" (UID: "88446f52-6434-4754-b225-3e79cbccac1b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.894899 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88446f52-6434-4754-b225-3e79cbccac1b" (UID: "88446f52-6434-4754-b225-3e79cbccac1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.898423 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "88446f52-6434-4754-b225-3e79cbccac1b" (UID: "88446f52-6434-4754-b225-3e79cbccac1b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.948950 4876 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.948998 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9gr9\" (UniqueName: \"kubernetes.io/projected/88446f52-6434-4754-b225-3e79cbccac1b-kube-api-access-n9gr9\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.949013 4876 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.949024 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.949035 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:22 crc kubenswrapper[4876]: I0313 07:59:22.965811 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-config-data" (OuterVolumeSpecName: "config-data") pod "88446f52-6434-4754-b225-3e79cbccac1b" (UID: "88446f52-6434-4754-b225-3e79cbccac1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.071065 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88446f52-6434-4754-b225-3e79cbccac1b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.193423 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.193535 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.243323 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dtrx5" event={"ID":"88446f52-6434-4754-b225-3e79cbccac1b","Type":"ContainerDied","Data":"66080dbd64209128231a8ea881031c7e752baa2ac26805332e1e84b53c583a7c"} Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.243378 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66080dbd64209128231a8ea881031c7e752baa2ac26805332e1e84b53c583a7c" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.243438 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dtrx5" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.795814 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.796333 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.898939 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.913581 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-556bf8599b-szn97"] Mar 13 07:59:23 crc kubenswrapper[4876]: E0313 07:59:23.914059 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88446f52-6434-4754-b225-3e79cbccac1b" containerName="keystone-bootstrap" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.914077 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="88446f52-6434-4754-b225-3e79cbccac1b" containerName="keystone-bootstrap" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.914292 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="88446f52-6434-4754-b225-3e79cbccac1b" containerName="keystone-bootstrap" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.915044 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.920966 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.921169 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.921287 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-btt8h" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.921407 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.921503 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.921589 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 07:59:23 crc kubenswrapper[4876]: I0313 07:59:23.927277 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-556bf8599b-szn97"] Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.096780 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-public-tls-certs\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.096849 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-config-data\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.096936 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-scripts\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.097137 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-fernet-keys\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.097374 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lscm7\" (UniqueName: \"kubernetes.io/projected/b9195731-b1ca-4c6c-90f3-54549fa3a813-kube-api-access-lscm7\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.097484 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-credential-keys\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.097508 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-internal-tls-certs\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.097536 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-combined-ca-bundle\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.200132 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lscm7\" (UniqueName: \"kubernetes.io/projected/b9195731-b1ca-4c6c-90f3-54549fa3a813-kube-api-access-lscm7\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.200213 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-credential-keys\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.200256 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-internal-tls-certs\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.200280 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-combined-ca-bundle\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.200325 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-public-tls-certs\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.200347 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-config-data\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.200372 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-scripts\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.200428 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-fernet-keys\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.208075 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-config-data\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.211729 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-scripts\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.216870 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-fernet-keys\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.217940 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-combined-ca-bundle\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.219946 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-public-tls-certs\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.220371 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-credential-keys\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.223895 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9195731-b1ca-4c6c-90f3-54549fa3a813-internal-tls-certs\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.228517 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lscm7\" (UniqueName: \"kubernetes.io/projected/b9195731-b1ca-4c6c-90f3-54549fa3a813-kube-api-access-lscm7\") pod \"keystone-556bf8599b-szn97\" (UID: \"b9195731-b1ca-4c6c-90f3-54549fa3a813\") " pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.269813 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.607616 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.608056 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.608112 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.640815 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8da8f1dc197e5903f65bce5f57eb064acd43a39afee0ba7b7e356d5bf1d3a13e"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 07:59:24 crc kubenswrapper[4876]: I0313 07:59:24.640937 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://8da8f1dc197e5903f65bce5f57eb064acd43a39afee0ba7b7e356d5bf1d3a13e" gracePeriod=600 Mar 13 07:59:25 crc kubenswrapper[4876]: I0313 07:59:25.017002 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 07:59:25 crc kubenswrapper[4876]: I0313 07:59:25.280847 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="8da8f1dc197e5903f65bce5f57eb064acd43a39afee0ba7b7e356d5bf1d3a13e" exitCode=0 Mar 13 07:59:25 crc kubenswrapper[4876]: I0313 07:59:25.280898 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"8da8f1dc197e5903f65bce5f57eb064acd43a39afee0ba7b7e356d5bf1d3a13e"} Mar 13 07:59:25 crc kubenswrapper[4876]: I0313 07:59:25.280941 4876 scope.go:117] "RemoveContainer" containerID="82566a3994e302f89937ff7ed8edf05f5c82529ba0688b6946e78bb6b8475402" Mar 13 07:59:25 crc kubenswrapper[4876]: I0313 07:59:25.328867 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: connect: connection refused" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.452837 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-l75wd" event={"ID":"669de0e1-d1a3-4298-bee2-e3713fd7f652","Type":"ContainerDied","Data":"fc5a33cf3379b1ddc5a013724c0ef1ab861099c497d927f6d741f0298b6570d3"} Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.453482 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc5a33cf3379b1ddc5a013724c0ef1ab861099c497d927f6d741f0298b6570d3" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.514518 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-l75wd" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.617217 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27g6x\" (UniqueName: \"kubernetes.io/projected/669de0e1-d1a3-4298-bee2-e3713fd7f652-kube-api-access-27g6x\") pod \"669de0e1-d1a3-4298-bee2-e3713fd7f652\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.618153 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/669de0e1-d1a3-4298-bee2-e3713fd7f652-logs\") pod \"669de0e1-d1a3-4298-bee2-e3713fd7f652\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.618286 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-combined-ca-bundle\") pod \"669de0e1-d1a3-4298-bee2-e3713fd7f652\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.618386 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-config-data\") pod \"669de0e1-d1a3-4298-bee2-e3713fd7f652\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.618433 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-scripts\") pod \"669de0e1-d1a3-4298-bee2-e3713fd7f652\" (UID: \"669de0e1-d1a3-4298-bee2-e3713fd7f652\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.619705 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/669de0e1-d1a3-4298-bee2-e3713fd7f652-logs" (OuterVolumeSpecName: "logs") pod "669de0e1-d1a3-4298-bee2-e3713fd7f652" (UID: "669de0e1-d1a3-4298-bee2-e3713fd7f652"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.626533 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-scripts" (OuterVolumeSpecName: "scripts") pod "669de0e1-d1a3-4298-bee2-e3713fd7f652" (UID: "669de0e1-d1a3-4298-bee2-e3713fd7f652"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.626672 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/669de0e1-d1a3-4298-bee2-e3713fd7f652-kube-api-access-27g6x" (OuterVolumeSpecName: "kube-api-access-27g6x") pod "669de0e1-d1a3-4298-bee2-e3713fd7f652" (UID: "669de0e1-d1a3-4298-bee2-e3713fd7f652"). InnerVolumeSpecName "kube-api-access-27g6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.668483 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-config-data" (OuterVolumeSpecName: "config-data") pod "669de0e1-d1a3-4298-bee2-e3713fd7f652" (UID: "669de0e1-d1a3-4298-bee2-e3713fd7f652"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.688947 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "669de0e1-d1a3-4298-bee2-e3713fd7f652" (UID: "669de0e1-d1a3-4298-bee2-e3713fd7f652"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.722338 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.722372 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.722384 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/669de0e1-d1a3-4298-bee2-e3713fd7f652-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.722393 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27g6x\" (UniqueName: \"kubernetes.io/projected/669de0e1-d1a3-4298-bee2-e3713fd7f652-kube-api-access-27g6x\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.722406 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/669de0e1-d1a3-4298-bee2-e3713fd7f652-logs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.734133 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66dd85954b-6789x" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.789271 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.824965 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-sb\") pod \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.825041 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-svc\") pod \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.825112 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-config\") pod \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.825192 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-swift-storage-0\") pod \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.825276 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqt2q\" (UniqueName: \"kubernetes.io/projected/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-kube-api-access-pqt2q\") pod \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.825332 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-nb\") pod \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\" (UID: \"2b6f7e51-08c6-4e1a-b099-31ffc41866ea\") " Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.888793 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-764bff8b56-4l8lg" podUID="70b62196-7476-4c85-a2d7-2299da069989" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.930139 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-kube-api-access-pqt2q" (OuterVolumeSpecName: "kube-api-access-pqt2q") pod "2b6f7e51-08c6-4e1a-b099-31ffc41866ea" (UID: "2b6f7e51-08c6-4e1a-b099-31ffc41866ea"). InnerVolumeSpecName "kube-api-access-pqt2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:28 crc kubenswrapper[4876]: I0313 07:59:28.930530 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqt2q\" (UniqueName: \"kubernetes.io/projected/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-kube-api-access-pqt2q\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.107600 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2b6f7e51-08c6-4e1a-b099-31ffc41866ea" (UID: "2b6f7e51-08c6-4e1a-b099-31ffc41866ea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.120137 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2b6f7e51-08c6-4e1a-b099-31ffc41866ea" (UID: "2b6f7e51-08c6-4e1a-b099-31ffc41866ea"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.131160 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-556bf8599b-szn97"] Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.135441 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.135473 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.164511 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b6f7e51-08c6-4e1a-b099-31ffc41866ea" (UID: "2b6f7e51-08c6-4e1a-b099-31ffc41866ea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.167490 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2b6f7e51-08c6-4e1a-b099-31ffc41866ea" (UID: "2b6f7e51-08c6-4e1a-b099-31ffc41866ea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.176434 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-config" (OuterVolumeSpecName: "config") pod "2b6f7e51-08c6-4e1a-b099-31ffc41866ea" (UID: "2b6f7e51-08c6-4e1a-b099-31ffc41866ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.243433 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.243480 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.243495 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b6f7e51-08c6-4e1a-b099-31ffc41866ea-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.479301 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" event={"ID":"2b6f7e51-08c6-4e1a-b099-31ffc41866ea","Type":"ContainerDied","Data":"8f694ac305e489193cf2a6a654ae1eeefb1dcd174dca1dec02ce1c8ec9d37098"} Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.479845 4876 scope.go:117] "RemoveContainer" containerID="36fff9f513672061d505049d00b22d172bdafe59341929dfa7826153aec5016e" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.479403 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5v9sw" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.487386 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rzbxc" event={"ID":"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32","Type":"ContainerStarted","Data":"fe3da91ed9a3a3ca316b34ea0f7c95a6a1766d17e9327d1fe2218db9cab2dd6c"} Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.493608 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerStarted","Data":"8065d8f3b387985421911bcc02b9ff1f0f13eca5ea11e59670b06fe74e46e7cd"} Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.494757 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-556bf8599b-szn97" event={"ID":"b9195731-b1ca-4c6c-90f3-54549fa3a813","Type":"ContainerStarted","Data":"0e3a2592dd615cc88c221262280fe3e18c13b1356c804f3b959959c4bceda1ec"} Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.494796 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-556bf8599b-szn97" event={"ID":"b9195731-b1ca-4c6c-90f3-54549fa3a813","Type":"ContainerStarted","Data":"976afc9f625300274c71e8ff8e164c01b1a1f90c8acc9c2217b2a649d2de1b7c"} Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.495403 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.503096 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-l75wd" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.503181 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"ab9fcd397cad6841be9000d6381846b64e56e1d04c705091fe8c6b9884bbef7d"} Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.549218 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-rzbxc" podStartSLOduration=3.355896612 podStartE2EDuration="50.549192442s" podCreationTimestamp="2026-03-13 07:58:39 +0000 UTC" firstStartedPulling="2026-03-13 07:58:41.403841509 +0000 UTC m=+1181.074620491" lastFinishedPulling="2026-03-13 07:59:28.597137339 +0000 UTC m=+1228.267916321" observedRunningTime="2026-03-13 07:59:29.517693267 +0000 UTC m=+1229.188472249" watchObservedRunningTime="2026-03-13 07:59:29.549192442 +0000 UTC m=+1229.219971424" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.572469 4876 scope.go:117] "RemoveContainer" containerID="e6471cb3f5e9fbe3c96869a454d8b4b0fedc8deb7f34dc47c585acb5bb24e3f8" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.577466 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-556bf8599b-szn97" podStartSLOduration=6.577434824 podStartE2EDuration="6.577434824s" podCreationTimestamp="2026-03-13 07:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:29.549154811 +0000 UTC m=+1229.219933793" watchObservedRunningTime="2026-03-13 07:59:29.577434824 +0000 UTC m=+1229.248213806" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.650032 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5v9sw"] Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.666801 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5v9sw"] Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.734414 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6dbd6d4d74-7zn9n"] Mar 13 07:59:29 crc kubenswrapper[4876]: E0313 07:59:29.734839 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="669de0e1-d1a3-4298-bee2-e3713fd7f652" containerName="placement-db-sync" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.734859 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="669de0e1-d1a3-4298-bee2-e3713fd7f652" containerName="placement-db-sync" Mar 13 07:59:29 crc kubenswrapper[4876]: E0313 07:59:29.734868 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerName="dnsmasq-dns" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.734875 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerName="dnsmasq-dns" Mar 13 07:59:29 crc kubenswrapper[4876]: E0313 07:59:29.734892 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerName="init" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.734898 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerName="init" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.735086 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" containerName="dnsmasq-dns" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.735106 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="669de0e1-d1a3-4298-bee2-e3713fd7f652" containerName="placement-db-sync" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.736043 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.739607 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n5vvq" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.739823 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.739929 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.739967 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.740396 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.754320 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-internal-tls-certs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.754375 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-scripts\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.754485 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ab4e2aa-c297-4979-9322-78f14c7b36f8-logs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.754527 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-config-data\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.754560 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-combined-ca-bundle\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.754594 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-public-tls-certs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.754689 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8fm4\" (UniqueName: \"kubernetes.io/projected/9ab4e2aa-c297-4979-9322-78f14c7b36f8-kube-api-access-f8fm4\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.767168 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6dbd6d4d74-7zn9n"] Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.856529 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-combined-ca-bundle\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.856596 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-public-tls-certs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.856674 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8fm4\" (UniqueName: \"kubernetes.io/projected/9ab4e2aa-c297-4979-9322-78f14c7b36f8-kube-api-access-f8fm4\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.856705 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-internal-tls-certs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.856727 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-scripts\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.856810 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ab4e2aa-c297-4979-9322-78f14c7b36f8-logs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.856844 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-config-data\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.868930 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ab4e2aa-c297-4979-9322-78f14c7b36f8-logs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.875954 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-internal-tls-certs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.885911 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-combined-ca-bundle\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.891004 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-public-tls-certs\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.891687 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-scripts\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.893902 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab4e2aa-c297-4979-9322-78f14c7b36f8-config-data\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:29 crc kubenswrapper[4876]: I0313 07:59:29.901938 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8fm4\" (UniqueName: \"kubernetes.io/projected/9ab4e2aa-c297-4979-9322-78f14c7b36f8-kube-api-access-f8fm4\") pod \"placement-6dbd6d4d74-7zn9n\" (UID: \"9ab4e2aa-c297-4979-9322-78f14c7b36f8\") " pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:30 crc kubenswrapper[4876]: I0313 07:59:30.059195 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:30 crc kubenswrapper[4876]: I0313 07:59:30.517284 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6dbd6d4d74-7zn9n"] Mar 13 07:59:30 crc kubenswrapper[4876]: I0313 07:59:30.524490 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ltwdh" event={"ID":"5d42486d-85b1-4384-91fc-10552a2c404b","Type":"ContainerStarted","Data":"33ad2db5455f1f79c44d2ce36ec5d0ed2b28712faf211c4126b8ffdd05dca8d7"} Mar 13 07:59:30 crc kubenswrapper[4876]: W0313 07:59:30.531677 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ab4e2aa_c297_4979_9322_78f14c7b36f8.slice/crio-01af2e23aa8c77a69c7ac8622e861c28805d59c5d0827d923778cad330e27ced WatchSource:0}: Error finding container 01af2e23aa8c77a69c7ac8622e861c28805d59c5d0827d923778cad330e27ced: Status 404 returned error can't find the container with id 01af2e23aa8c77a69c7ac8622e861c28805d59c5d0827d923778cad330e27ced Mar 13 07:59:30 crc kubenswrapper[4876]: I0313 07:59:30.566578 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-ltwdh" podStartSLOduration=3.636290029 podStartE2EDuration="51.566550981s" podCreationTimestamp="2026-03-13 07:58:39 +0000 UTC" firstStartedPulling="2026-03-13 07:58:40.670177902 +0000 UTC m=+1180.340956884" lastFinishedPulling="2026-03-13 07:59:28.600438854 +0000 UTC m=+1228.271217836" observedRunningTime="2026-03-13 07:59:30.558594174 +0000 UTC m=+1230.229373156" watchObservedRunningTime="2026-03-13 07:59:30.566550981 +0000 UTC m=+1230.237329963" Mar 13 07:59:31 crc kubenswrapper[4876]: I0313 07:59:31.052566 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b6f7e51-08c6-4e1a-b099-31ffc41866ea" path="/var/lib/kubelet/pods/2b6f7e51-08c6-4e1a-b099-31ffc41866ea/volumes" Mar 13 07:59:31 crc kubenswrapper[4876]: I0313 07:59:31.574686 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbd6d4d74-7zn9n" event={"ID":"9ab4e2aa-c297-4979-9322-78f14c7b36f8","Type":"ContainerStarted","Data":"9bdcc2a7377b537171f5c03685c7afc875a66c91aa9853ddab4f7f1c979b6be0"} Mar 13 07:59:31 crc kubenswrapper[4876]: I0313 07:59:31.575123 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbd6d4d74-7zn9n" event={"ID":"9ab4e2aa-c297-4979-9322-78f14c7b36f8","Type":"ContainerStarted","Data":"01af2e23aa8c77a69c7ac8622e861c28805d59c5d0827d923778cad330e27ced"} Mar 13 07:59:32 crc kubenswrapper[4876]: I0313 07:59:32.592200 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dbd6d4d74-7zn9n" event={"ID":"9ab4e2aa-c297-4979-9322-78f14c7b36f8","Type":"ContainerStarted","Data":"31f5cf3d33c85a197cea06e68cc619873718e6d7860dad79f71a125e9dcb5bb5"} Mar 13 07:59:32 crc kubenswrapper[4876]: I0313 07:59:32.592552 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:32 crc kubenswrapper[4876]: I0313 07:59:32.592579 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 07:59:32 crc kubenswrapper[4876]: I0313 07:59:32.905322 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6dbd6d4d74-7zn9n" podStartSLOduration=3.905296494 podStartE2EDuration="3.905296494s" podCreationTimestamp="2026-03-13 07:59:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:32.901045865 +0000 UTC m=+1232.571824847" watchObservedRunningTime="2026-03-13 07:59:32.905296494 +0000 UTC m=+1232.576075476" Mar 13 07:59:34 crc kubenswrapper[4876]: I0313 07:59:34.622234 4876 generic.go:334] "Generic (PLEG): container finished" podID="86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" containerID="fe3da91ed9a3a3ca316b34ea0f7c95a6a1766d17e9327d1fe2218db9cab2dd6c" exitCode=0 Mar 13 07:59:34 crc kubenswrapper[4876]: I0313 07:59:34.622371 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rzbxc" event={"ID":"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32","Type":"ContainerDied","Data":"fe3da91ed9a3a3ca316b34ea0f7c95a6a1766d17e9327d1fe2218db9cab2dd6c"} Mar 13 07:59:37 crc kubenswrapper[4876]: I0313 07:59:37.662873 4876 generic.go:334] "Generic (PLEG): container finished" podID="5d42486d-85b1-4384-91fc-10552a2c404b" containerID="33ad2db5455f1f79c44d2ce36ec5d0ed2b28712faf211c4126b8ffdd05dca8d7" exitCode=0 Mar 13 07:59:37 crc kubenswrapper[4876]: I0313 07:59:37.662997 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ltwdh" event={"ID":"5d42486d-85b1-4384-91fc-10552a2c404b","Type":"ContainerDied","Data":"33ad2db5455f1f79c44d2ce36ec5d0ed2b28712faf211c4126b8ffdd05dca8d7"} Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.084982 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.272467 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tndxt\" (UniqueName: \"kubernetes.io/projected/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-kube-api-access-tndxt\") pod \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.272519 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-db-sync-config-data\") pod \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.272599 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-combined-ca-bundle\") pod \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\" (UID: \"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32\") " Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.278254 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" (UID: "86d9e4f3-c1e8-40fe-a6b8-79507f63ca32"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.279604 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-kube-api-access-tndxt" (OuterVolumeSpecName: "kube-api-access-tndxt") pod "86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" (UID: "86d9e4f3-c1e8-40fe-a6b8-79507f63ca32"). InnerVolumeSpecName "kube-api-access-tndxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.365783 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" (UID: "86d9e4f3-c1e8-40fe-a6b8-79507f63ca32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.374780 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tndxt\" (UniqueName: \"kubernetes.io/projected/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-kube-api-access-tndxt\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.374817 4876 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.374829 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.674859 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rzbxc" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.674839 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rzbxc" event={"ID":"86d9e4f3-c1e8-40fe-a6b8-79507f63ca32","Type":"ContainerDied","Data":"9cec9de6c021b4153884f9cca0464e4b3e939357dc84f7d753dec6c76994a476"} Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.675299 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cec9de6c021b4153884f9cca0464e4b3e939357dc84f7d753dec6c76994a476" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.716828 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66dd85954b-6789x" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.823056 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-764bff8b56-4l8lg" podUID="70b62196-7476-4c85-a2d7-2299da069989" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 07:59:38 crc kubenswrapper[4876]: I0313 07:59:38.946872 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.098224 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-scripts\") pod \"5d42486d-85b1-4384-91fc-10552a2c404b\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.098353 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d42486d-85b1-4384-91fc-10552a2c404b-etc-machine-id\") pod \"5d42486d-85b1-4384-91fc-10552a2c404b\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.098392 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-db-sync-config-data\") pod \"5d42486d-85b1-4384-91fc-10552a2c404b\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.098508 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-combined-ca-bundle\") pod \"5d42486d-85b1-4384-91fc-10552a2c404b\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.098501 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d42486d-85b1-4384-91fc-10552a2c404b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5d42486d-85b1-4384-91fc-10552a2c404b" (UID: "5d42486d-85b1-4384-91fc-10552a2c404b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.098628 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-config-data\") pod \"5d42486d-85b1-4384-91fc-10552a2c404b\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.098694 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz79q\" (UniqueName: \"kubernetes.io/projected/5d42486d-85b1-4384-91fc-10552a2c404b-kube-api-access-sz79q\") pod \"5d42486d-85b1-4384-91fc-10552a2c404b\" (UID: \"5d42486d-85b1-4384-91fc-10552a2c404b\") " Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.099110 4876 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d42486d-85b1-4384-91fc-10552a2c404b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.107429 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5d42486d-85b1-4384-91fc-10552a2c404b" (UID: "5d42486d-85b1-4384-91fc-10552a2c404b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.108624 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-scripts" (OuterVolumeSpecName: "scripts") pod "5d42486d-85b1-4384-91fc-10552a2c404b" (UID: "5d42486d-85b1-4384-91fc-10552a2c404b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.110130 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d42486d-85b1-4384-91fc-10552a2c404b-kube-api-access-sz79q" (OuterVolumeSpecName: "kube-api-access-sz79q") pod "5d42486d-85b1-4384-91fc-10552a2c404b" (UID: "5d42486d-85b1-4384-91fc-10552a2c404b"). InnerVolumeSpecName "kube-api-access-sz79q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.176014 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-config-data" (OuterVolumeSpecName: "config-data") pod "5d42486d-85b1-4384-91fc-10552a2c404b" (UID: "5d42486d-85b1-4384-91fc-10552a2c404b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.197804 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d42486d-85b1-4384-91fc-10552a2c404b" (UID: "5d42486d-85b1-4384-91fc-10552a2c404b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.202408 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz79q\" (UniqueName: \"kubernetes.io/projected/5d42486d-85b1-4384-91fc-10552a2c404b-kube-api-access-sz79q\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.202458 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.202473 4876 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.202483 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.202492 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42486d-85b1-4384-91fc-10552a2c404b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.413434 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5864c7c4b6-qqb8c"] Mar 13 07:59:39 crc kubenswrapper[4876]: E0313 07:59:39.413926 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" containerName="barbican-db-sync" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.413945 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" containerName="barbican-db-sync" Mar 13 07:59:39 crc kubenswrapper[4876]: E0313 07:59:39.413968 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d42486d-85b1-4384-91fc-10552a2c404b" containerName="cinder-db-sync" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.413975 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d42486d-85b1-4384-91fc-10552a2c404b" containerName="cinder-db-sync" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.414172 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d42486d-85b1-4384-91fc-10552a2c404b" containerName="cinder-db-sync" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.414195 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" containerName="barbican-db-sync" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.415299 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.426586 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.426964 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-p54ct" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.427210 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.463858 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5864c7c4b6-qqb8c"] Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.520540 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5467d6f7c5-djg95"] Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.523344 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5467d6f7c5-djg95"] Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.523533 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.528486 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rl69\" (UniqueName: \"kubernetes.io/projected/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-kube-api-access-9rl69\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.528673 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-config-data\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.528752 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-logs\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.528837 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-combined-ca-bundle\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.528950 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-config-data-custom\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.529442 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.631829 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94630a52-4ec8-4ba9-966b-334ec46e9d25-logs\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.631930 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rl69\" (UniqueName: \"kubernetes.io/projected/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-kube-api-access-9rl69\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.631956 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-config-data-custom\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.631985 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-config-data\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.632022 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-combined-ca-bundle\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.632044 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-config-data\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.632065 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-logs\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.632093 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-combined-ca-bundle\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.632116 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-config-data-custom\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.632157 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx8qb\" (UniqueName: \"kubernetes.io/projected/94630a52-4ec8-4ba9-966b-334ec46e9d25-kube-api-access-vx8qb\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.643933 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-logs\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.664095 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-z8md7"] Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.678983 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.719986 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rl69\" (UniqueName: \"kubernetes.io/projected/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-kube-api-access-9rl69\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.737785 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-combined-ca-bundle\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.737989 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94630a52-4ec8-4ba9-966b-334ec46e9d25-logs\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.738293 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-config-data-custom\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.738425 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-config-data\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.738566 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-combined-ca-bundle\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.738797 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx8qb\" (UniqueName: \"kubernetes.io/projected/94630a52-4ec8-4ba9-966b-334ec46e9d25-kube-api-access-vx8qb\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.739484 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94630a52-4ec8-4ba9-966b-334ec46e9d25-logs\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.740488 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-config-data\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.744911 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-z8md7"] Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.747290 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-config-data-custom\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.748821 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79f28ab7-204b-4e1c-be09-9e48b9ed5a75-config-data-custom\") pod \"barbican-keystone-listener-5864c7c4b6-qqb8c\" (UID: \"79f28ab7-204b-4e1c-be09-9e48b9ed5a75\") " pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.750029 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.753040 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-config-data\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.757428 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94630a52-4ec8-4ba9-966b-334ec46e9d25-combined-ca-bundle\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.804041 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ltwdh" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.815043 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx8qb\" (UniqueName: \"kubernetes.io/projected/94630a52-4ec8-4ba9-966b-334ec46e9d25-kube-api-access-vx8qb\") pod \"barbican-worker-5467d6f7c5-djg95\" (UID: \"94630a52-4ec8-4ba9-966b-334ec46e9d25\") " pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.815418 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ltwdh" event={"ID":"5d42486d-85b1-4384-91fc-10552a2c404b","Type":"ContainerDied","Data":"98b1fb16c85acd3567987e5ddc33472b976ee850ad6261f8278fbb918351bcbf"} Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.815470 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98b1fb16c85acd3567987e5ddc33472b976ee850ad6261f8278fbb918351bcbf" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.844019 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-svc\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.844092 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbj9s\" (UniqueName: \"kubernetes.io/projected/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-kube-api-access-cbj9s\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.844136 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.844203 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.844222 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.844259 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-config\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.869463 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerStarted","Data":"bba9f685500d5103dcfef11e9fbf6d51e86619e17f8245e274c3009e4a681e8a"} Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.869700 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="ceilometer-central-agent" containerID="cri-o://2e99ade3cdbc5460384ad6c59b414c3d139f3f8b1523d2abf7cd47152566ea2a" gracePeriod=30 Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.870075 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.870415 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="proxy-httpd" containerID="cri-o://bba9f685500d5103dcfef11e9fbf6d51e86619e17f8245e274c3009e4a681e8a" gracePeriod=30 Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.870469 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="sg-core" containerID="cri-o://8065d8f3b387985421911bcc02b9ff1f0f13eca5ea11e59670b06fe74e46e7cd" gracePeriod=30 Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.870514 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="ceilometer-notification-agent" containerID="cri-o://d21cc25bd945c08dee67e81d7642e373e6e49cc67322be5d3e60e5c28f6d5bf5" gracePeriod=30 Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.908909 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5467d6f7c5-djg95" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.947217 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-svc\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.947304 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbj9s\" (UniqueName: \"kubernetes.io/projected/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-kube-api-access-cbj9s\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.947344 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.947404 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.947421 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.947441 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-config\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.948466 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-config\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.954386 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.955628 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-858fd4b7b-stb59"] Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.958625 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-svc\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.959638 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:39 crc kubenswrapper[4876]: E0313 07:59:39.963375 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d42486d_85b1_4384_91fc_10552a2c404b.slice\": RecentStats: unable to find data in memory cache]" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.979123 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.986454 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 13 07:59:39 crc kubenswrapper[4876]: I0313 07:59:39.955072 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.010755 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-858fd4b7b-stb59"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.013192 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbj9s\" (UniqueName: \"kubernetes.io/projected/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-kube-api-access-cbj9s\") pod \"dnsmasq-dns-85ff748b95-z8md7\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.058181 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.082356 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.500495731 podStartE2EDuration="1m1.082332018s" podCreationTimestamp="2026-03-13 07:58:39 +0000 UTC" firstStartedPulling="2026-03-13 07:58:41.13078383 +0000 UTC m=+1180.801562812" lastFinishedPulling="2026-03-13 07:59:38.712620117 +0000 UTC m=+1238.383399099" observedRunningTime="2026-03-13 07:59:40.042754512 +0000 UTC m=+1239.713533494" watchObservedRunningTime="2026-03-13 07:59:40.082332018 +0000 UTC m=+1239.753111000" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.087727 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.168426 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.168867 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a2479e9-5256-4e71-8f79-44cebfb76fa7-logs\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.168959 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data-custom\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.169011 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4vjj\" (UniqueName: \"kubernetes.io/projected/4a2479e9-5256-4e71-8f79-44cebfb76fa7-kube-api-access-w4vjj\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.169031 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-combined-ca-bundle\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.275572 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.276929 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.279049 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a2479e9-5256-4e71-8f79-44cebfb76fa7-logs\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.279168 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.281805 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data-custom\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.282093 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4vjj\" (UniqueName: \"kubernetes.io/projected/4a2479e9-5256-4e71-8f79-44cebfb76fa7-kube-api-access-w4vjj\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.282153 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-combined-ca-bundle\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.299666 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a2479e9-5256-4e71-8f79-44cebfb76fa7-logs\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.309160 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.309190 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.309416 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5gdvr" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.309459 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.341504 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-combined-ca-bundle\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.352628 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.362630 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4vjj\" (UniqueName: \"kubernetes.io/projected/4a2479e9-5256-4e71-8f79-44cebfb76fa7-kube-api-access-w4vjj\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.362735 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data-custom\") pod \"barbican-api-858fd4b7b-stb59\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.375033 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.384446 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-scripts\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.384558 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87dqh\" (UniqueName: \"kubernetes.io/projected/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-kube-api-access-87dqh\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.384598 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.384630 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.384704 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.384783 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.457511 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-z8md7"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.490504 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.490688 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-scripts\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.490719 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87dqh\" (UniqueName: \"kubernetes.io/projected/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-kube-api-access-87dqh\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.490779 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.490802 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.490872 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.504993 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-7rqrl"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.506332 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.506418 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.509797 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.513521 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-scripts\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.513824 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.517950 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.521909 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87dqh\" (UniqueName: \"kubernetes.io/projected/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-kube-api-access-87dqh\") pod \"cinder-scheduler-0\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.528857 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-7rqrl"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.613283 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.675273 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.721521 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.760442 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.760550 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.760619 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.760678 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l95pl\" (UniqueName: \"kubernetes.io/projected/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-kube-api-access-l95pl\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.760817 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.760893 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-config\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.792911 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.798713 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.798903 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.809552 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78dc766d77-2lq9j"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.809920 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-78dc766d77-2lq9j" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-api" containerID="cri-o://80661cfa70c99dd7b3d101dbcd3d9f87ed202b141eb93e53b24e735b71649c7c" gracePeriod=30 Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.810733 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-78dc766d77-2lq9j" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-httpd" containerID="cri-o://a82353ca03963734c9e8d399b80dc22325e344e78ee20af877d27c28fbfe3390" gracePeriod=30 Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.831927 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5884ccb8d9-4wjtm"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.837499 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.845083 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5884ccb8d9-4wjtm"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866472 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866526 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-logs\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866568 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866602 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866638 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l95pl\" (UniqueName: \"kubernetes.io/projected/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-kube-api-access-l95pl\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866684 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866710 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-scripts\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866729 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr4jt\" (UniqueName: \"kubernetes.io/projected/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-kube-api-access-mr4jt\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866746 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866772 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866802 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data-custom\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866823 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-config\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.866853 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.868585 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.869443 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5864c7c4b6-qqb8c"] Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.870504 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.871468 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.874724 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.892584 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-config\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.918731 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l95pl\" (UniqueName: \"kubernetes.io/projected/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-kube-api-access-l95pl\") pod \"dnsmasq-dns-5c9776ccc5-7rqrl\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.972664 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-public-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.972985 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-logs\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973027 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-internal-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973065 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-httpd-config\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973085 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-combined-ca-bundle\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973310 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-scripts\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973347 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr4jt\" (UniqueName: \"kubernetes.io/projected/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-kube-api-access-mr4jt\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973370 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-ovndb-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973388 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973419 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973452 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data-custom\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973474 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-config\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973509 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.973534 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht2wx\" (UniqueName: \"kubernetes.io/projected/cfccefa5-a3f2-4513-a79b-372dd903ac2d-kube-api-access-ht2wx\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.974074 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-logs\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.980133 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-scripts\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.984295 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.985250 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.985949 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:40 crc kubenswrapper[4876]: I0313 07:59:40.987853 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data-custom\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.009079 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr4jt\" (UniqueName: \"kubernetes.io/projected/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-kube-api-access-mr4jt\") pod \"cinder-api-0\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " pod="openstack/cinder-api-0" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.080404 4876 generic.go:334] "Generic (PLEG): container finished" podID="c287c63f-40c2-427d-a58e-fea0297c947b" containerID="bba9f685500d5103dcfef11e9fbf6d51e86619e17f8245e274c3009e4a681e8a" exitCode=0 Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.080459 4876 generic.go:334] "Generic (PLEG): container finished" podID="c287c63f-40c2-427d-a58e-fea0297c947b" containerID="8065d8f3b387985421911bcc02b9ff1f0f13eca5ea11e59670b06fe74e46e7cd" exitCode=2 Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.080929 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-ovndb-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.081038 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-config\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.081073 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht2wx\" (UniqueName: \"kubernetes.io/projected/cfccefa5-a3f2-4513-a79b-372dd903ac2d-kube-api-access-ht2wx\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.081110 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-public-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.081160 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-internal-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.081185 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-httpd-config\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.081204 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-combined-ca-bundle\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.095205 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-combined-ca-bundle\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.096099 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-ovndb-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.113990 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-public-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.119207 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-config\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.123699 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-internal-tls-certs\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.126689 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerDied","Data":"bba9f685500d5103dcfef11e9fbf6d51e86619e17f8245e274c3009e4a681e8a"} Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.126759 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerDied","Data":"8065d8f3b387985421911bcc02b9ff1f0f13eca5ea11e59670b06fe74e46e7cd"} Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.126771 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" event={"ID":"79f28ab7-204b-4e1c-be09-9e48b9ed5a75","Type":"ContainerStarted","Data":"eaaa46117680ae4c2d4ea91f7e007c51d09a1a17ae16fbd0877f42bf6726ed81"} Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.128935 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cfccefa5-a3f2-4513-a79b-372dd903ac2d-httpd-config\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.129297 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht2wx\" (UniqueName: \"kubernetes.io/projected/cfccefa5-a3f2-4513-a79b-372dd903ac2d-kube-api-access-ht2wx\") pod \"neutron-5884ccb8d9-4wjtm\" (UID: \"cfccefa5-a3f2-4513-a79b-372dd903ac2d\") " pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.165302 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-z8md7"] Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.166408 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.188108 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.189664 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.192383 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5467d6f7c5-djg95"] Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.198586 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-78dc766d77-2lq9j" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9696/\": read tcp 10.217.0.2:36048->10.217.0.158:9696: read: connection reset by peer" Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.444010 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.509000 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-858fd4b7b-stb59"] Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.923290 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-7rqrl"] Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.938095 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5884ccb8d9-4wjtm"] Mar 13 07:59:41 crc kubenswrapper[4876]: I0313 07:59:41.950453 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 07:59:41 crc kubenswrapper[4876]: W0313 07:59:41.967099 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfccefa5_a3f2_4513_a79b_372dd903ac2d.slice/crio-9ef999854ead0186d39a3cc6894c3dd71f68e0b2e24a4d229b7d777e22f06be9 WatchSource:0}: Error finding container 9ef999854ead0186d39a3cc6894c3dd71f68e0b2e24a4d229b7d777e22f06be9: Status 404 returned error can't find the container with id 9ef999854ead0186d39a3cc6894c3dd71f68e0b2e24a4d229b7d777e22f06be9 Mar 13 07:59:41 crc kubenswrapper[4876]: W0313 07:59:41.990483 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c0e0e38_68d2_45a0_9601_fe2e4f3f9179.slice/crio-aca84be5eed54930d8c202168ed96f5d816f3847da9538b4df3f7e179eff33d9 WatchSource:0}: Error finding container aca84be5eed54930d8c202168ed96f5d816f3847da9538b4df3f7e179eff33d9: Status 404 returned error can't find the container with id aca84be5eed54930d8c202168ed96f5d816f3847da9538b4df3f7e179eff33d9 Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.152421 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179","Type":"ContainerStarted","Data":"aca84be5eed54930d8c202168ed96f5d816f3847da9538b4df3f7e179eff33d9"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.157530 4876 generic.go:334] "Generic (PLEG): container finished" podID="c287c63f-40c2-427d-a58e-fea0297c947b" containerID="2e99ade3cdbc5460384ad6c59b414c3d139f3f8b1523d2abf7cd47152566ea2a" exitCode=0 Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.157589 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerDied","Data":"2e99ade3cdbc5460384ad6c59b414c3d139f3f8b1523d2abf7cd47152566ea2a"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.158976 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1","Type":"ContainerStarted","Data":"9f58792e93f9cfd15343e710842c7195ec034b093cbc3ea1fdccb4d2f8eff8e3"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.161486 4876 generic.go:334] "Generic (PLEG): container finished" podID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerID="a82353ca03963734c9e8d399b80dc22325e344e78ee20af877d27c28fbfe3390" exitCode=0 Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.161527 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc766d77-2lq9j" event={"ID":"130c3a6d-4650-4cb3-a46b-9a08eb9108e3","Type":"ContainerDied","Data":"a82353ca03963734c9e8d399b80dc22325e344e78ee20af877d27c28fbfe3390"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.165004 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-858fd4b7b-stb59" event={"ID":"4a2479e9-5256-4e71-8f79-44cebfb76fa7","Type":"ContainerStarted","Data":"df53d39702dc0583b7cefe144f4f08804c1dfa962ef426cd072075b277e9921e"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.165032 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-858fd4b7b-stb59" event={"ID":"4a2479e9-5256-4e71-8f79-44cebfb76fa7","Type":"ContainerStarted","Data":"f97161a4ebe6b77286166276d77ab64a2abee0cb4e24e1191be8df52c6cbb398"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.170191 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5467d6f7c5-djg95" event={"ID":"94630a52-4ec8-4ba9-966b-334ec46e9d25","Type":"ContainerStarted","Data":"36aa564c62e82f9e2f8b71f0ae7bd208e4813a061ae19b4d391c10b0c7653b80"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.175523 4876 generic.go:334] "Generic (PLEG): container finished" podID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerID="0bebe7f9d85929a662bbc5ed96fc28b1d498412a36ff797ea0962b227bd9884d" exitCode=137 Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.175550 4876 generic.go:334] "Generic (PLEG): container finished" podID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerID="289009ab13035d7e67f3fef975563573dcbd7b63473eda841713899882a68394" exitCode=137 Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.175590 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59996cc657-lxnh9" event={"ID":"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a","Type":"ContainerDied","Data":"0bebe7f9d85929a662bbc5ed96fc28b1d498412a36ff797ea0962b227bd9884d"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.175641 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59996cc657-lxnh9" event={"ID":"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a","Type":"ContainerDied","Data":"289009ab13035d7e67f3fef975563573dcbd7b63473eda841713899882a68394"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.191101 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5884ccb8d9-4wjtm" event={"ID":"cfccefa5-a3f2-4513-a79b-372dd903ac2d","Type":"ContainerStarted","Data":"9ef999854ead0186d39a3cc6894c3dd71f68e0b2e24a4d229b7d777e22f06be9"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.194210 4876 generic.go:334] "Generic (PLEG): container finished" podID="77ea0d81-d539-4ccf-9280-98ba2f77ffc4" containerID="8894aa4f85279d09acd15fc0d197c863bf3135ff03572334180dd504ffd1e85d" exitCode=0 Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.194279 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-z8md7" event={"ID":"77ea0d81-d539-4ccf-9280-98ba2f77ffc4","Type":"ContainerDied","Data":"8894aa4f85279d09acd15fc0d197c863bf3135ff03572334180dd504ffd1e85d"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.194303 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-z8md7" event={"ID":"77ea0d81-d539-4ccf-9280-98ba2f77ffc4","Type":"ContainerStarted","Data":"00471e868154a60d0acce217fb803fda4d4031347e84836a19998f48948b0460"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.197662 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" event={"ID":"43cdaeee-26ec-47e3-88e9-eda01a1dde0b","Type":"ContainerStarted","Data":"c43e0ee356b7f1d208345facb16858c6a34eca75fd97df3f51ac8a1ab2da29e6"} Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.573427 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.690973 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-logs\") pod \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.693099 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-logs" (OuterVolumeSpecName: "logs") pod "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" (UID: "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.712381 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-scripts\") pod \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.712559 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xntrc\" (UniqueName: \"kubernetes.io/projected/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-kube-api-access-xntrc\") pod \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.712605 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-horizon-secret-key\") pod \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.712917 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-config-data\") pod \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\" (UID: \"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a\") " Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.744408 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" (UID: "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.765255 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-kube-api-access-xntrc" (OuterVolumeSpecName: "kube-api-access-xntrc") pod "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" (UID: "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a"). InnerVolumeSpecName "kube-api-access-xntrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.781046 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xntrc\" (UniqueName: \"kubernetes.io/projected/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-kube-api-access-xntrc\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.781089 4876 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.781112 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-logs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.909055 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-scripts" (OuterVolumeSpecName: "scripts") pod "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" (UID: "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.914417 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-config-data" (OuterVolumeSpecName: "config-data") pod "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" (UID: "4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.990620 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:42 crc kubenswrapper[4876]: I0313 07:59:42.991138 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.097431 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-78dc766d77-2lq9j" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9696/\": dial tcp 10.217.0.158:9696: connect: connection refused" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.271022 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-858fd4b7b-stb59" event={"ID":"4a2479e9-5256-4e71-8f79-44cebfb76fa7","Type":"ContainerStarted","Data":"fe205b2df6edfb12f85b95b867ce4fc59c121522b1f5725082c986daf4eb8f74"} Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.272472 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.272546 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.281954 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-z8md7" event={"ID":"77ea0d81-d539-4ccf-9280-98ba2f77ffc4","Type":"ContainerDied","Data":"00471e868154a60d0acce217fb803fda4d4031347e84836a19998f48948b0460"} Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.282004 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00471e868154a60d0acce217fb803fda4d4031347e84836a19998f48948b0460" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.290645 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59996cc657-lxnh9" event={"ID":"4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a","Type":"ContainerDied","Data":"4a3147076ee9e24ee442e33b726dc0146d6b7647b44bb3da55a3e46d011ede22"} Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.290709 4876 scope.go:117] "RemoveContainer" containerID="0bebe7f9d85929a662bbc5ed96fc28b1d498412a36ff797ea0962b227bd9884d" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.290821 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59996cc657-lxnh9" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.314426 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-858fd4b7b-stb59" podStartSLOduration=4.314402576 podStartE2EDuration="4.314402576s" podCreationTimestamp="2026-03-13 07:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:43.302680948 +0000 UTC m=+1242.973459930" watchObservedRunningTime="2026-03-13 07:59:43.314402576 +0000 UTC m=+1242.985181558" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.314590 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5884ccb8d9-4wjtm" event={"ID":"cfccefa5-a3f2-4513-a79b-372dd903ac2d","Type":"ContainerStarted","Data":"e0d1b7707118b960d767ce3f776726d5702d0f8ebed5b4d7399a4e75bf20b34c"} Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.363774 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.428219 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-config\") pod \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.428296 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-nb\") pod \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.428346 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-svc\") pod \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.428415 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-swift-storage-0\") pod \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.428471 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbj9s\" (UniqueName: \"kubernetes.io/projected/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-kube-api-access-cbj9s\") pod \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.428561 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-sb\") pod \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\" (UID: \"77ea0d81-d539-4ccf-9280-98ba2f77ffc4\") " Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.446860 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59996cc657-lxnh9"] Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.468149 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-59996cc657-lxnh9"] Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.500770 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-kube-api-access-cbj9s" (OuterVolumeSpecName: "kube-api-access-cbj9s") pod "77ea0d81-d539-4ccf-9280-98ba2f77ffc4" (UID: "77ea0d81-d539-4ccf-9280-98ba2f77ffc4"). InnerVolumeSpecName "kube-api-access-cbj9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.532025 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbj9s\" (UniqueName: \"kubernetes.io/projected/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-kube-api-access-cbj9s\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.610060 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-config" (OuterVolumeSpecName: "config") pod "77ea0d81-d539-4ccf-9280-98ba2f77ffc4" (UID: "77ea0d81-d539-4ccf-9280-98ba2f77ffc4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.628673 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "77ea0d81-d539-4ccf-9280-98ba2f77ffc4" (UID: "77ea0d81-d539-4ccf-9280-98ba2f77ffc4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.636045 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.636099 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.696839 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "77ea0d81-d539-4ccf-9280-98ba2f77ffc4" (UID: "77ea0d81-d539-4ccf-9280-98ba2f77ffc4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.699414 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "77ea0d81-d539-4ccf-9280-98ba2f77ffc4" (UID: "77ea0d81-d539-4ccf-9280-98ba2f77ffc4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.704846 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "77ea0d81-d539-4ccf-9280-98ba2f77ffc4" (UID: "77ea0d81-d539-4ccf-9280-98ba2f77ffc4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.739262 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.739763 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:43 crc kubenswrapper[4876]: I0313 07:59:43.739774 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77ea0d81-d539-4ccf-9280-98ba2f77ffc4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.255959 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.326514 4876 generic.go:334] "Generic (PLEG): container finished" podID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerID="8b0b01e6fbcbef1c8dbf931b5f260df83af53d917f840ae99c4e7e5295d70609" exitCode=0 Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.326615 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" event={"ID":"43cdaeee-26ec-47e3-88e9-eda01a1dde0b","Type":"ContainerDied","Data":"8b0b01e6fbcbef1c8dbf931b5f260df83af53d917f840ae99c4e7e5295d70609"} Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.331764 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179","Type":"ContainerStarted","Data":"a98aa8db7fa9f9b7ee2f43daf4ea074a3f58593cf14ae19768bbde77177f49bb"} Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.336484 4876 generic.go:334] "Generic (PLEG): container finished" podID="c287c63f-40c2-427d-a58e-fea0297c947b" containerID="d21cc25bd945c08dee67e81d7642e373e6e49cc67322be5d3e60e5c28f6d5bf5" exitCode=0 Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.336552 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerDied","Data":"d21cc25bd945c08dee67e81d7642e373e6e49cc67322be5d3e60e5c28f6d5bf5"} Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.338351 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1","Type":"ContainerStarted","Data":"f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d"} Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.340477 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-z8md7" Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.344830 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5884ccb8d9-4wjtm" event={"ID":"cfccefa5-a3f2-4513-a79b-372dd903ac2d","Type":"ContainerStarted","Data":"3364795637e610f3cf854ebd6af7aa04ffad2ae62b3744d102ed602156bbd808"} Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.346213 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.384929 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5884ccb8d9-4wjtm" podStartSLOduration=4.384893806 podStartE2EDuration="4.384893806s" podCreationTimestamp="2026-03-13 07:59:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:44.376741638 +0000 UTC m=+1244.047520620" watchObservedRunningTime="2026-03-13 07:59:44.384893806 +0000 UTC m=+1244.055672788" Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.450337 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-z8md7"] Mar 13 07:59:44 crc kubenswrapper[4876]: I0313 07:59:44.471616 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-z8md7"] Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.070856 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" path="/var/lib/kubelet/pods/4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a/volumes" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.072223 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77ea0d81-d539-4ccf-9280-98ba2f77ffc4" path="/var/lib/kubelet/pods/77ea0d81-d539-4ccf-9280-98ba2f77ffc4/volumes" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.089076 4876 scope.go:117] "RemoveContainer" containerID="289009ab13035d7e67f3fef975563573dcbd7b63473eda841713899882a68394" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.447005 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c287c63f-40c2-427d-a58e-fea0297c947b","Type":"ContainerDied","Data":"d896f9559f6684b16d084597175cba274598b862a29dcbcddd07b3232f18d142"} Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.447546 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d896f9559f6684b16d084597175cba274598b862a29dcbcddd07b3232f18d142" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.447784 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.713155 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-run-httpd\") pod \"c287c63f-40c2-427d-a58e-fea0297c947b\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.713208 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-log-httpd\") pod \"c287c63f-40c2-427d-a58e-fea0297c947b\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.713248 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-scripts\") pod \"c287c63f-40c2-427d-a58e-fea0297c947b\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.713446 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-combined-ca-bundle\") pod \"c287c63f-40c2-427d-a58e-fea0297c947b\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.713499 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb8qj\" (UniqueName: \"kubernetes.io/projected/c287c63f-40c2-427d-a58e-fea0297c947b-kube-api-access-sb8qj\") pod \"c287c63f-40c2-427d-a58e-fea0297c947b\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.713551 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-config-data\") pod \"c287c63f-40c2-427d-a58e-fea0297c947b\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.713609 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-sg-core-conf-yaml\") pod \"c287c63f-40c2-427d-a58e-fea0297c947b\" (UID: \"c287c63f-40c2-427d-a58e-fea0297c947b\") " Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.721848 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c287c63f-40c2-427d-a58e-fea0297c947b" (UID: "c287c63f-40c2-427d-a58e-fea0297c947b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.722382 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c287c63f-40c2-427d-a58e-fea0297c947b" (UID: "c287c63f-40c2-427d-a58e-fea0297c947b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.727832 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c287c63f-40c2-427d-a58e-fea0297c947b-kube-api-access-sb8qj" (OuterVolumeSpecName: "kube-api-access-sb8qj") pod "c287c63f-40c2-427d-a58e-fea0297c947b" (UID: "c287c63f-40c2-427d-a58e-fea0297c947b"). InnerVolumeSpecName "kube-api-access-sb8qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.733468 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-scripts" (OuterVolumeSpecName: "scripts") pod "c287c63f-40c2-427d-a58e-fea0297c947b" (UID: "c287c63f-40c2-427d-a58e-fea0297c947b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.818181 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.818212 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c287c63f-40c2-427d-a58e-fea0297c947b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.818246 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.818257 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb8qj\" (UniqueName: \"kubernetes.io/projected/c287c63f-40c2-427d-a58e-fea0297c947b-kube-api-access-sb8qj\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.823157 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c287c63f-40c2-427d-a58e-fea0297c947b" (UID: "c287c63f-40c2-427d-a58e-fea0297c947b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.920405 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.920580 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c287c63f-40c2-427d-a58e-fea0297c947b" (UID: "c287c63f-40c2-427d-a58e-fea0297c947b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:45 crc kubenswrapper[4876]: I0313 07:59:45.934422 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-config-data" (OuterVolumeSpecName: "config-data") pod "c287c63f-40c2-427d-a58e-fea0297c947b" (UID: "c287c63f-40c2-427d-a58e-fea0297c947b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.023121 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.023162 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c287c63f-40c2-427d-a58e-fea0297c947b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.509649 4876 generic.go:334] "Generic (PLEG): container finished" podID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerID="80661cfa70c99dd7b3d101dbcd3d9f87ed202b141eb93e53b24e735b71649c7c" exitCode=0 Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.510344 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc766d77-2lq9j" event={"ID":"130c3a6d-4650-4cb3-a46b-9a08eb9108e3","Type":"ContainerDied","Data":"80661cfa70c99dd7b3d101dbcd3d9f87ed202b141eb93e53b24e735b71649c7c"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.510701 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-78dc766d77-2lq9j" event={"ID":"130c3a6d-4650-4cb3-a46b-9a08eb9108e3","Type":"ContainerDied","Data":"4da2c3a76a1815831dbf68cac85263b6ac0549a3f22ddc8d30e4de7d3dd63210"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.510729 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4da2c3a76a1815831dbf68cac85263b6ac0549a3f22ddc8d30e4de7d3dd63210" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.521748 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5467d6f7c5-djg95" event={"ID":"94630a52-4ec8-4ba9-966b-334ec46e9d25","Type":"ContainerStarted","Data":"c14e339e1d83f58527f9c4ad7c3d1d8192246f9115a302cc39482b782b7920b4"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.521820 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5467d6f7c5-djg95" event={"ID":"94630a52-4ec8-4ba9-966b-334ec46e9d25","Type":"ContainerStarted","Data":"0cbc0c21ab7bfdcd077f9a5473adc8df6a9464d248090162e05bc48d8cd17cfc"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.528162 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179","Type":"ContainerStarted","Data":"f759362668103ca973b7087882f219d00529c8d9e78138c3c9d8b0b0451038b6"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.528371 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.528361 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerName="cinder-api" containerID="cri-o://f759362668103ca973b7087882f219d00529c8d9e78138c3c9d8b0b0451038b6" gracePeriod=30 Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.528504 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerName="cinder-api-log" containerID="cri-o://a98aa8db7fa9f9b7ee2f43daf4ea074a3f58593cf14ae19768bbde77177f49bb" gracePeriod=30 Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.537445 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" event={"ID":"43cdaeee-26ec-47e3-88e9-eda01a1dde0b","Type":"ContainerStarted","Data":"148e33290be9b0f22e06f90a2510aab995fd0af4f7d3acd9c95fa7efa28dd22b"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.537561 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.548130 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5467d6f7c5-djg95" podStartSLOduration=3.6415491060000003 podStartE2EDuration="7.548101629s" podCreationTimestamp="2026-03-13 07:59:39 +0000 UTC" firstStartedPulling="2026-03-13 07:59:41.199904404 +0000 UTC m=+1240.870683376" lastFinishedPulling="2026-03-13 07:59:45.106456917 +0000 UTC m=+1244.777235899" observedRunningTime="2026-03-13 07:59:46.546032181 +0000 UTC m=+1246.216811163" watchObservedRunningTime="2026-03-13 07:59:46.548101629 +0000 UTC m=+1246.218880611" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.550838 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" event={"ID":"79f28ab7-204b-4e1c-be09-9e48b9ed5a75","Type":"ContainerStarted","Data":"366c2849e7eff1229edbb8daa7d9443619d7b64d9d951eff3a30d56888a767d0"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.550902 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" event={"ID":"79f28ab7-204b-4e1c-be09-9e48b9ed5a75","Type":"ContainerStarted","Data":"24e93da284c736598f754d6893000004c31aa6bdbd2a33fbcbe652fef2e1ab99"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.566596 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.569400 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1","Type":"ContainerStarted","Data":"5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174"} Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.569761 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.598572 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.598548838 podStartE2EDuration="6.598548838s" podCreationTimestamp="2026-03-13 07:59:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:46.587861859 +0000 UTC m=+1246.258640851" watchObservedRunningTime="2026-03-13 07:59:46.598548838 +0000 UTC m=+1246.269327830" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.611751 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" podStartSLOduration=6.611716756 podStartE2EDuration="6.611716756s" podCreationTimestamp="2026-03-13 07:59:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:46.606229683 +0000 UTC m=+1246.277008665" watchObservedRunningTime="2026-03-13 07:59:46.611716756 +0000 UTC m=+1246.282495738" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.657931 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-public-tls-certs\") pod \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.658056 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-combined-ca-bundle\") pod \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.658139 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-internal-tls-certs\") pod \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.666617 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-994tn\" (UniqueName: \"kubernetes.io/projected/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-kube-api-access-994tn\") pod \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.666695 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-config\") pod \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.666757 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-httpd-config\") pod \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.666797 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-ovndb-tls-certs\") pod \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\" (UID: \"130c3a6d-4650-4cb3-a46b-9a08eb9108e3\") " Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.674111 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.082493188 podStartE2EDuration="6.674083509s" podCreationTimestamp="2026-03-13 07:59:40 +0000 UTC" firstStartedPulling="2026-03-13 07:59:41.461918415 +0000 UTC m=+1241.132697397" lastFinishedPulling="2026-03-13 07:59:43.053508736 +0000 UTC m=+1242.724287718" observedRunningTime="2026-03-13 07:59:46.659514342 +0000 UTC m=+1246.330293324" watchObservedRunningTime="2026-03-13 07:59:46.674083509 +0000 UTC m=+1246.344862491" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.716725 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5864c7c4b6-qqb8c" podStartSLOduration=3.426201918 podStartE2EDuration="7.716692429s" podCreationTimestamp="2026-03-13 07:59:39 +0000 UTC" firstStartedPulling="2026-03-13 07:59:40.814151366 +0000 UTC m=+1240.484930348" lastFinishedPulling="2026-03-13 07:59:45.104641877 +0000 UTC m=+1244.775420859" observedRunningTime="2026-03-13 07:59:46.680994082 +0000 UTC m=+1246.351773064" watchObservedRunningTime="2026-03-13 07:59:46.716692429 +0000 UTC m=+1246.387471411" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.739638 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-kube-api-access-994tn" (OuterVolumeSpecName: "kube-api-access-994tn") pod "130c3a6d-4650-4cb3-a46b-9a08eb9108e3" (UID: "130c3a6d-4650-4cb3-a46b-9a08eb9108e3"). InnerVolumeSpecName "kube-api-access-994tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.763609 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "130c3a6d-4650-4cb3-a46b-9a08eb9108e3" (UID: "130c3a6d-4650-4cb3-a46b-9a08eb9108e3"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.765812 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.769753 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-994tn\" (UniqueName: \"kubernetes.io/projected/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-kube-api-access-994tn\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.769788 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.781388 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.791633 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792114 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="proxy-httpd" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792141 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="proxy-httpd" Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792152 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77ea0d81-d539-4ccf-9280-98ba2f77ffc4" containerName="init" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792158 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="77ea0d81-d539-4ccf-9280-98ba2f77ffc4" containerName="init" Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792178 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="ceilometer-notification-agent" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792185 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="ceilometer-notification-agent" Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792196 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-httpd" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792202 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-httpd" Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792213 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-api" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792218 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-api" Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792228 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerName="horizon" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792249 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerName="horizon" Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792276 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="sg-core" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792285 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="sg-core" Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792296 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="ceilometer-central-agent" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792302 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="ceilometer-central-agent" Mar 13 07:59:46 crc kubenswrapper[4876]: E0313 07:59:46.792311 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerName="horizon-log" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792317 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerName="horizon-log" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792492 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-api" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792504 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="77ea0d81-d539-4ccf-9280-98ba2f77ffc4" containerName="init" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792517 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="sg-core" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792527 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" containerName="neutron-httpd" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792538 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="ceilometer-notification-agent" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792550 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerName="horizon-log" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792560 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a" containerName="horizon" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792568 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="proxy-httpd" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.792578 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" containerName="ceilometer-central-agent" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.794457 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.797874 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-config" (OuterVolumeSpecName: "config") pod "130c3a6d-4650-4cb3-a46b-9a08eb9108e3" (UID: "130c3a6d-4650-4cb3-a46b-9a08eb9108e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.798792 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.798944 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.805547 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.811004 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "130c3a6d-4650-4cb3-a46b-9a08eb9108e3" (UID: "130c3a6d-4650-4cb3-a46b-9a08eb9108e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.852382 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "130c3a6d-4650-4cb3-a46b-9a08eb9108e3" (UID: "130c3a6d-4650-4cb3-a46b-9a08eb9108e3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.864627 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "130c3a6d-4650-4cb3-a46b-9a08eb9108e3" (UID: "130c3a6d-4650-4cb3-a46b-9a08eb9108e3"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872109 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-config-data\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872181 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872212 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-run-httpd\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872443 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kmlp\" (UniqueName: \"kubernetes.io/projected/f43b9345-cbf5-4885-9ac4-8e944d8d3142-kube-api-access-2kmlp\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872489 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-log-httpd\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872522 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-scripts\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872563 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872636 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872650 4876 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872664 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.872675 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.891828 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "130c3a6d-4650-4cb3-a46b-9a08eb9108e3" (UID: "130c3a6d-4650-4cb3-a46b-9a08eb9108e3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.974993 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kmlp\" (UniqueName: \"kubernetes.io/projected/f43b9345-cbf5-4885-9ac4-8e944d8d3142-kube-api-access-2kmlp\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.975098 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-log-httpd\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.975147 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-scripts\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.975200 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.975281 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-config-data\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.975326 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.975366 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-run-httpd\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.975464 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/130c3a6d-4650-4cb3-a46b-9a08eb9108e3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.976063 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-run-httpd\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.976896 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-log-httpd\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.981630 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.983777 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-scripts\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.990900 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.991487 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-config-data\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:46 crc kubenswrapper[4876]: I0313 07:59:46.999107 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kmlp\" (UniqueName: \"kubernetes.io/projected/f43b9345-cbf5-4885-9ac4-8e944d8d3142-kube-api-access-2kmlp\") pod \"ceilometer-0\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " pod="openstack/ceilometer-0" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.049259 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c287c63f-40c2-427d-a58e-fea0297c947b" path="/var/lib/kubelet/pods/c287c63f-40c2-427d-a58e-fea0297c947b/volumes" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.127066 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.201881 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8b86bd47d-2hv8s"] Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.203586 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.213526 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.213829 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.235798 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8b86bd47d-2hv8s"] Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.285005 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-config-data\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.285213 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-combined-ca-bundle\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.285340 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-logs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.285363 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-internal-tls-certs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.285399 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl97m\" (UniqueName: \"kubernetes.io/projected/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-kube-api-access-wl97m\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.285446 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-config-data-custom\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.285462 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-public-tls-certs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.387460 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-config-data-custom\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.387524 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-public-tls-certs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.387566 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-config-data\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.387595 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-combined-ca-bundle\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.387701 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-logs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.387735 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-internal-tls-certs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.387775 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl97m\" (UniqueName: \"kubernetes.io/projected/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-kube-api-access-wl97m\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.390560 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-logs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.397840 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-config-data\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.415705 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-combined-ca-bundle\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.422249 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-config-data-custom\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.423374 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-internal-tls-certs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.423372 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-public-tls-certs\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.427872 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl97m\" (UniqueName: \"kubernetes.io/projected/65a5d87e-b9ef-4be3-ade6-46fa03282d7d-kube-api-access-wl97m\") pod \"barbican-api-8b86bd47d-2hv8s\" (UID: \"65a5d87e-b9ef-4be3-ade6-46fa03282d7d\") " pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.584844 4876 generic.go:334] "Generic (PLEG): container finished" podID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerID="a98aa8db7fa9f9b7ee2f43daf4ea074a3f58593cf14ae19768bbde77177f49bb" exitCode=143 Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.585750 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179","Type":"ContainerDied","Data":"a98aa8db7fa9f9b7ee2f43daf4ea074a3f58593cf14ae19768bbde77177f49bb"} Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.585806 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-78dc766d77-2lq9j" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.611474 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.690682 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-78dc766d77-2lq9j"] Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.709194 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-78dc766d77-2lq9j"] Mar 13 07:59:47 crc kubenswrapper[4876]: I0313 07:59:47.807755 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 07:59:48 crc kubenswrapper[4876]: I0313 07:59:48.145812 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8b86bd47d-2hv8s"] Mar 13 07:59:48 crc kubenswrapper[4876]: I0313 07:59:48.596919 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerStarted","Data":"87ada5e68ea0261af0a10cb1d652a149eaead4ad156ccf9bea4c23ae75fa8b7a"} Mar 13 07:59:48 crc kubenswrapper[4876]: I0313 07:59:48.597972 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8b86bd47d-2hv8s" event={"ID":"65a5d87e-b9ef-4be3-ade6-46fa03282d7d","Type":"ContainerStarted","Data":"c7cae7e40ee49ff97cec553bad27389592da5eca0e4e9303c498372ae5a935c3"} Mar 13 07:59:49 crc kubenswrapper[4876]: I0313 07:59:49.046694 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130c3a6d-4650-4cb3-a46b-9a08eb9108e3" path="/var/lib/kubelet/pods/130c3a6d-4650-4cb3-a46b-9a08eb9108e3/volumes" Mar 13 07:59:49 crc kubenswrapper[4876]: I0313 07:59:49.612697 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerStarted","Data":"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621"} Mar 13 07:59:49 crc kubenswrapper[4876]: I0313 07:59:49.618311 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8b86bd47d-2hv8s" event={"ID":"65a5d87e-b9ef-4be3-ade6-46fa03282d7d","Type":"ContainerStarted","Data":"41ec092253c31ab1d35fef9ff3a720e5709000dabc8ce16f62fc9e5d73c44069"} Mar 13 07:59:49 crc kubenswrapper[4876]: I0313 07:59:49.618870 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8b86bd47d-2hv8s" event={"ID":"65a5d87e-b9ef-4be3-ade6-46fa03282d7d","Type":"ContainerStarted","Data":"7e5e5c0a923d54a34cf6735f628471b1c1a5eb9fdf7402c4b48df12e24d342b8"} Mar 13 07:59:49 crc kubenswrapper[4876]: I0313 07:59:49.619000 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:49 crc kubenswrapper[4876]: I0313 07:59:49.619023 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 07:59:49 crc kubenswrapper[4876]: I0313 07:59:49.649947 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8b86bd47d-2hv8s" podStartSLOduration=2.649926646 podStartE2EDuration="2.649926646s" podCreationTimestamp="2026-03-13 07:59:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 07:59:49.6471887 +0000 UTC m=+1249.317967692" watchObservedRunningTime="2026-03-13 07:59:49.649926646 +0000 UTC m=+1249.320705628" Mar 13 07:59:50 crc kubenswrapper[4876]: I0313 07:59:50.634340 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerStarted","Data":"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3"} Mar 13 07:59:50 crc kubenswrapper[4876]: I0313 07:59:50.634909 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerStarted","Data":"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553"} Mar 13 07:59:50 crc kubenswrapper[4876]: I0313 07:59:50.678163 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.077624 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.163224 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.190532 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.274988 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-cv8h6"] Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.275724 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" podUID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" containerName="dnsmasq-dns" containerID="cri-o://374ddce7f5025c68c1d9607cb69dd15de70e48ec3b5796bf7fa40ff95677c903" gracePeriod=10 Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.415833 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.663496 4876 generic.go:334] "Generic (PLEG): container finished" podID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" containerID="374ddce7f5025c68c1d9607cb69dd15de70e48ec3b5796bf7fa40ff95677c903" exitCode=0 Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.663987 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" event={"ID":"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1","Type":"ContainerDied","Data":"374ddce7f5025c68c1d9607cb69dd15de70e48ec3b5796bf7fa40ff95677c903"} Mar 13 07:59:51 crc kubenswrapper[4876]: I0313 07:59:51.742591 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.031584 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.167059 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-svc\") pod \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.167137 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-config\") pod \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.167197 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-swift-storage-0\") pod \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.167270 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-sb\") pod \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.167431 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fxjd\" (UniqueName: \"kubernetes.io/projected/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-kube-api-access-8fxjd\") pod \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.167462 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-nb\") pod \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\" (UID: \"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1\") " Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.191468 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-kube-api-access-8fxjd" (OuterVolumeSpecName: "kube-api-access-8fxjd") pod "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" (UID: "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1"). InnerVolumeSpecName "kube-api-access-8fxjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.249804 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" (UID: "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.271079 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fxjd\" (UniqueName: \"kubernetes.io/projected/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-kube-api-access-8fxjd\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.271129 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.279066 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" (UID: "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.279217 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" (UID: "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.296067 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-config" (OuterVolumeSpecName: "config") pod "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" (UID: "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.338158 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" (UID: "6b7e85e2-cb9b-4d19-87df-f611b0abc5e1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.373063 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.373114 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-config\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.373124 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.373134 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.680542 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.680538 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-cv8h6" event={"ID":"6b7e85e2-cb9b-4d19-87df-f611b0abc5e1","Type":"ContainerDied","Data":"32a43100d669dd65512a75a5eb896be277937b64ac0913ce1613a22bce1c84a9"} Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.681147 4876 scope.go:117] "RemoveContainer" containerID="374ddce7f5025c68c1d9607cb69dd15de70e48ec3b5796bf7fa40ff95677c903" Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.680714 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerName="cinder-scheduler" containerID="cri-o://f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d" gracePeriod=30 Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.680798 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerName="probe" containerID="cri-o://5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174" gracePeriod=30 Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.733483 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-cv8h6"] Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.743308 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-cv8h6"] Mar 13 07:59:52 crc kubenswrapper[4876]: I0313 07:59:52.760484 4876 scope.go:117] "RemoveContainer" containerID="f38dea91a9d754ab4d9689a4afc91a181cdc068b29cd15f1a6e47c49200cb625" Mar 13 07:59:53 crc kubenswrapper[4876]: I0313 07:59:53.100414 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" path="/var/lib/kubelet/pods/6b7e85e2-cb9b-4d19-87df-f611b0abc5e1/volumes" Mar 13 07:59:53 crc kubenswrapper[4876]: I0313 07:59:53.215991 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:53 crc kubenswrapper[4876]: I0313 07:59:53.697349 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerStarted","Data":"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec"} Mar 13 07:59:53 crc kubenswrapper[4876]: I0313 07:59:53.697446 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 07:59:53 crc kubenswrapper[4876]: I0313 07:59:53.701672 4876 generic.go:334] "Generic (PLEG): container finished" podID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerID="5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174" exitCode=0 Mar 13 07:59:53 crc kubenswrapper[4876]: I0313 07:59:53.701738 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1","Type":"ContainerDied","Data":"5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174"} Mar 13 07:59:53 crc kubenswrapper[4876]: I0313 07:59:53.722383 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.075180447 podStartE2EDuration="7.722364315s" podCreationTimestamp="2026-03-13 07:59:46 +0000 UTC" firstStartedPulling="2026-03-13 07:59:47.81012175 +0000 UTC m=+1247.480900732" lastFinishedPulling="2026-03-13 07:59:52.457305618 +0000 UTC m=+1252.128084600" observedRunningTime="2026-03-13 07:59:53.720107312 +0000 UTC m=+1253.390886294" watchObservedRunningTime="2026-03-13 07:59:53.722364315 +0000 UTC m=+1253.393143297" Mar 13 07:59:54 crc kubenswrapper[4876]: I0313 07:59:54.127417 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-66dd85954b-6789x" Mar 13 07:59:54 crc kubenswrapper[4876]: I0313 07:59:54.316926 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 07:59:54 crc kubenswrapper[4876]: I0313 07:59:54.553748 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-764bff8b56-4l8lg" Mar 13 07:59:54 crc kubenswrapper[4876]: I0313 07:59:54.630948 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66dd85954b-6789x"] Mar 13 07:59:54 crc kubenswrapper[4876]: I0313 07:59:54.713384 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66dd85954b-6789x" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon-log" containerID="cri-o://b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8" gracePeriod=30 Mar 13 07:59:54 crc kubenswrapper[4876]: I0313 07:59:54.713630 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66dd85954b-6789x" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" containerID="cri-o://645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74" gracePeriod=30 Mar 13 07:59:54 crc kubenswrapper[4876]: I0313 07:59:54.827772 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 07:59:56 crc kubenswrapper[4876]: I0313 07:59:56.529910 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-556bf8599b-szn97" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.514427 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.617404 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data\") pod \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.617522 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-etc-machine-id\") pod \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.617608 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87dqh\" (UniqueName: \"kubernetes.io/projected/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-kube-api-access-87dqh\") pod \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.617636 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data-custom\") pod \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.617795 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-scripts\") pod \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.617845 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-combined-ca-bundle\") pod \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\" (UID: \"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1\") " Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.620375 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" (UID: "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.626547 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-kube-api-access-87dqh" (OuterVolumeSpecName: "kube-api-access-87dqh") pod "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" (UID: "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1"). InnerVolumeSpecName "kube-api-access-87dqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.627697 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-scripts" (OuterVolumeSpecName: "scripts") pod "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" (UID: "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.660463 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" (UID: "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.709115 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" (UID: "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.721763 4876 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.721803 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87dqh\" (UniqueName: \"kubernetes.io/projected/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-kube-api-access-87dqh\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.721817 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.721826 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.721835 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.768579 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data" (OuterVolumeSpecName: "config-data") pod "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" (UID: "81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.811343 4876 generic.go:334] "Generic (PLEG): container finished" podID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerID="f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d" exitCode=0 Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.811417 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1","Type":"ContainerDied","Data":"f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d"} Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.811469 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1","Type":"ContainerDied","Data":"9f58792e93f9cfd15343e710842c7195ec034b093cbc3ea1fdccb4d2f8eff8e3"} Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.811496 4876 scope.go:117] "RemoveContainer" containerID="5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.811499 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.823512 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.856545 4876 scope.go:117] "RemoveContainer" containerID="f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.869005 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.884598 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.890195 4876 scope.go:117] "RemoveContainer" containerID="5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174" Mar 13 07:59:57 crc kubenswrapper[4876]: E0313 07:59:57.892358 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174\": container with ID starting with 5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174 not found: ID does not exist" containerID="5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.892413 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174"} err="failed to get container status \"5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174\": rpc error: code = NotFound desc = could not find container \"5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174\": container with ID starting with 5bd82633c8b5e18631dd12cedd519e7c8acb472efa3932f9ecf2dfedbe148174 not found: ID does not exist" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.892439 4876 scope.go:117] "RemoveContainer" containerID="f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d" Mar 13 07:59:57 crc kubenswrapper[4876]: E0313 07:59:57.896474 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d\": container with ID starting with f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d not found: ID does not exist" containerID="f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.896548 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d"} err="failed to get container status \"f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d\": rpc error: code = NotFound desc = could not find container \"f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d\": container with ID starting with f6840b778f823b4890700a7739cbcbde32425d56e4a0ac034b8c89c61f27c17d not found: ID does not exist" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.899476 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:57 crc kubenswrapper[4876]: E0313 07:59:57.900004 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerName="cinder-scheduler" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.900026 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerName="cinder-scheduler" Mar 13 07:59:57 crc kubenswrapper[4876]: E0313 07:59:57.900047 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" containerName="init" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.900055 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" containerName="init" Mar 13 07:59:57 crc kubenswrapper[4876]: E0313 07:59:57.900074 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" containerName="dnsmasq-dns" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.900080 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" containerName="dnsmasq-dns" Mar 13 07:59:57 crc kubenswrapper[4876]: E0313 07:59:57.900096 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerName="probe" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.900102 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerName="probe" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.900304 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerName="probe" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.900324 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" containerName="cinder-scheduler" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.900343 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b7e85e2-cb9b-4d19-87df-f611b0abc5e1" containerName="dnsmasq-dns" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.901468 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.905108 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.914573 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.925378 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.925438 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.925473 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.925676 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c3cb708-df89-4a10-ba84-b0c55efd2b61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.925806 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwn6g\" (UniqueName: \"kubernetes.io/projected/4c3cb708-df89-4a10-ba84-b0c55efd2b61-kube-api-access-nwn6g\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:57 crc kubenswrapper[4876]: I0313 07:59:57.925843 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.027628 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c3cb708-df89-4a10-ba84-b0c55efd2b61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.027745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwn6g\" (UniqueName: \"kubernetes.io/projected/4c3cb708-df89-4a10-ba84-b0c55efd2b61-kube-api-access-nwn6g\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.027782 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.027829 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.027858 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.027887 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.028881 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c3cb708-df89-4a10-ba84-b0c55efd2b61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.035642 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.039178 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.041166 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-config-data\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.067169 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c3cb708-df89-4a10-ba84-b0c55efd2b61-scripts\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.078025 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwn6g\" (UniqueName: \"kubernetes.io/projected/4c3cb708-df89-4a10-ba84-b0c55efd2b61-kube-api-access-nwn6g\") pod \"cinder-scheduler-0\" (UID: \"4c3cb708-df89-4a10-ba84-b0c55efd2b61\") " pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.234306 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.721504 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66dd85954b-6789x" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.826328 4876 generic.go:334] "Generic (PLEG): container finished" podID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerID="645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74" exitCode=0 Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.826381 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66dd85954b-6789x" event={"ID":"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163","Type":"ContainerDied","Data":"645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74"} Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.895667 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.896921 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.899751 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.900148 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.900353 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-dwzfr" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.908069 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.920854 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.949369 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf9bw\" (UniqueName: \"kubernetes.io/projected/02e40021-e864-4c53-9522-40fe5719abde-kube-api-access-wf9bw\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.949477 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02e40021-e864-4c53-9522-40fe5719abde-openstack-config-secret\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.949508 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e40021-e864-4c53-9522-40fe5719abde-combined-ca-bundle\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:58 crc kubenswrapper[4876]: I0313 07:59:58.949574 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02e40021-e864-4c53-9522-40fe5719abde-openstack-config\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.051368 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf9bw\" (UniqueName: \"kubernetes.io/projected/02e40021-e864-4c53-9522-40fe5719abde-kube-api-access-wf9bw\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.051492 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02e40021-e864-4c53-9522-40fe5719abde-openstack-config-secret\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.051524 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e40021-e864-4c53-9522-40fe5719abde-combined-ca-bundle\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.051608 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02e40021-e864-4c53-9522-40fe5719abde-openstack-config\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.052523 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02e40021-e864-4c53-9522-40fe5719abde-openstack-config\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.057907 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02e40021-e864-4c53-9522-40fe5719abde-openstack-config-secret\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.064039 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e40021-e864-4c53-9522-40fe5719abde-combined-ca-bundle\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.079731 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf9bw\" (UniqueName: \"kubernetes.io/projected/02e40021-e864-4c53-9522-40fe5719abde-kube-api-access-wf9bw\") pod \"openstackclient\" (UID: \"02e40021-e864-4c53-9522-40fe5719abde\") " pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.106924 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1" path="/var/lib/kubelet/pods/81a3bdf1-3422-41b8-ab42-7b3fd6d59fd1/volumes" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.231954 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.661735 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.854864 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"02e40021-e864-4c53-9522-40fe5719abde","Type":"ContainerStarted","Data":"a8a3f731cf842eb39fd66c5afee44722a3908d6ce22798136ce35b195903eca9"} Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.902342 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c3cb708-df89-4a10-ba84-b0c55efd2b61","Type":"ContainerStarted","Data":"dd80e531211056fffb9c25b7479678d80386889467ef3f54550e412174ab1856"} Mar 13 07:59:59 crc kubenswrapper[4876]: I0313 07:59:59.902424 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c3cb708-df89-4a10-ba84-b0c55efd2b61","Type":"ContainerStarted","Data":"409181afebfda1fbab7d46afdfe8c7907d90f776ba6413ade29c7ef908269e32"} Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.209454 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd"] Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.211423 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.223932 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.224109 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.228372 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556480-8292c"] Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.229898 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556480-8292c" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.238989 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd"] Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.248094 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.248436 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.248694 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.258460 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556480-8292c"] Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.304080 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lptf9\" (UniqueName: \"kubernetes.io/projected/306b6d6a-7ec0-41f2-b7e6-8974bb873700-kube-api-access-lptf9\") pod \"auto-csr-approver-29556480-8292c\" (UID: \"306b6d6a-7ec0-41f2-b7e6-8974bb873700\") " pod="openshift-infra/auto-csr-approver-29556480-8292c" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.407600 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556fa6a5-90ac-43ea-a365-abd7ea03f21a-secret-volume\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.407665 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjl8s\" (UniqueName: \"kubernetes.io/projected/556fa6a5-90ac-43ea-a365-abd7ea03f21a-kube-api-access-pjl8s\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.407706 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lptf9\" (UniqueName: \"kubernetes.io/projected/306b6d6a-7ec0-41f2-b7e6-8974bb873700-kube-api-access-lptf9\") pod \"auto-csr-approver-29556480-8292c\" (UID: \"306b6d6a-7ec0-41f2-b7e6-8974bb873700\") " pod="openshift-infra/auto-csr-approver-29556480-8292c" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.407729 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556fa6a5-90ac-43ea-a365-abd7ea03f21a-config-volume\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.435079 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lptf9\" (UniqueName: \"kubernetes.io/projected/306b6d6a-7ec0-41f2-b7e6-8974bb873700-kube-api-access-lptf9\") pod \"auto-csr-approver-29556480-8292c\" (UID: \"306b6d6a-7ec0-41f2-b7e6-8974bb873700\") " pod="openshift-infra/auto-csr-approver-29556480-8292c" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.511799 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556fa6a5-90ac-43ea-a365-abd7ea03f21a-config-volume\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.512034 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556fa6a5-90ac-43ea-a365-abd7ea03f21a-secret-volume\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.512079 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjl8s\" (UniqueName: \"kubernetes.io/projected/556fa6a5-90ac-43ea-a365-abd7ea03f21a-kube-api-access-pjl8s\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.513901 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556fa6a5-90ac-43ea-a365-abd7ea03f21a-config-volume\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.520375 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556fa6a5-90ac-43ea-a365-abd7ea03f21a-secret-volume\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.555548 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjl8s\" (UniqueName: \"kubernetes.io/projected/556fa6a5-90ac-43ea-a365-abd7ea03f21a-kube-api-access-pjl8s\") pod \"collect-profiles-29556480-q2pqd\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.586773 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.604852 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556480-8292c" Mar 13 08:00:00 crc kubenswrapper[4876]: I0313 08:00:00.994768 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4c3cb708-df89-4a10-ba84-b0c55efd2b61","Type":"ContainerStarted","Data":"6cd053b5b8409067c21fd5894b6d11da3a027b6701dbc90281ee56dcf320594b"} Mar 13 08:00:01 crc kubenswrapper[4876]: I0313 08:00:01.086011 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.085969952 podStartE2EDuration="4.085969952s" podCreationTimestamp="2026-03-13 07:59:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:01.023863377 +0000 UTC m=+1260.694642369" watchObservedRunningTime="2026-03-13 08:00:01.085969952 +0000 UTC m=+1260.756748934" Mar 13 08:00:01 crc kubenswrapper[4876]: I0313 08:00:01.306747 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 08:00:01 crc kubenswrapper[4876]: I0313 08:00:01.319551 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd"] Mar 13 08:00:01 crc kubenswrapper[4876]: I0313 08:00:01.334578 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8b86bd47d-2hv8s" Mar 13 08:00:01 crc kubenswrapper[4876]: I0313 08:00:01.498551 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-858fd4b7b-stb59"] Mar 13 08:00:01 crc kubenswrapper[4876]: I0313 08:00:01.498920 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-858fd4b7b-stb59" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerName="barbican-api-log" containerID="cri-o://df53d39702dc0583b7cefe144f4f08804c1dfa962ef426cd072075b277e9921e" gracePeriod=30 Mar 13 08:00:01 crc kubenswrapper[4876]: I0313 08:00:01.499604 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-858fd4b7b-stb59" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerName="barbican-api" containerID="cri-o://fe205b2df6edfb12f85b95b867ce4fc59c121522b1f5725082c986daf4eb8f74" gracePeriod=30 Mar 13 08:00:01 crc kubenswrapper[4876]: I0313 08:00:01.668839 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556480-8292c"] Mar 13 08:00:02 crc kubenswrapper[4876]: I0313 08:00:02.050657 4876 generic.go:334] "Generic (PLEG): container finished" podID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerID="df53d39702dc0583b7cefe144f4f08804c1dfa962ef426cd072075b277e9921e" exitCode=143 Mar 13 08:00:02 crc kubenswrapper[4876]: I0313 08:00:02.051184 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-858fd4b7b-stb59" event={"ID":"4a2479e9-5256-4e71-8f79-44cebfb76fa7","Type":"ContainerDied","Data":"df53d39702dc0583b7cefe144f4f08804c1dfa962ef426cd072075b277e9921e"} Mar 13 08:00:02 crc kubenswrapper[4876]: I0313 08:00:02.073955 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556480-8292c" event={"ID":"306b6d6a-7ec0-41f2-b7e6-8974bb873700","Type":"ContainerStarted","Data":"ce54e944e19f28d93ba865a5d9e7939cbcd27d4e386d9c7760f5ec68eb51242c"} Mar 13 08:00:02 crc kubenswrapper[4876]: I0313 08:00:02.093542 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" event={"ID":"556fa6a5-90ac-43ea-a365-abd7ea03f21a","Type":"ContainerStarted","Data":"f98f31af3ff8ab23b2d9f95916001dd27779a009e580066cf5aafb24a665bc37"} Mar 13 08:00:02 crc kubenswrapper[4876]: I0313 08:00:02.093600 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" event={"ID":"556fa6a5-90ac-43ea-a365-abd7ea03f21a","Type":"ContainerStarted","Data":"fc7bb76035c345d0e092b1f79701c4c42304c805e7a77fc98167b3dfe6b720c3"} Mar 13 08:00:02 crc kubenswrapper[4876]: I0313 08:00:02.141450 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" podStartSLOduration=2.141430033 podStartE2EDuration="2.141430033s" podCreationTimestamp="2026-03-13 08:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:02.127888434 +0000 UTC m=+1261.798667416" watchObservedRunningTime="2026-03-13 08:00:02.141430033 +0000 UTC m=+1261.812209015" Mar 13 08:00:03 crc kubenswrapper[4876]: I0313 08:00:03.108693 4876 generic.go:334] "Generic (PLEG): container finished" podID="556fa6a5-90ac-43ea-a365-abd7ea03f21a" containerID="f98f31af3ff8ab23b2d9f95916001dd27779a009e580066cf5aafb24a665bc37" exitCode=0 Mar 13 08:00:03 crc kubenswrapper[4876]: I0313 08:00:03.108799 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" event={"ID":"556fa6a5-90ac-43ea-a365-abd7ea03f21a","Type":"ContainerDied","Data":"f98f31af3ff8ab23b2d9f95916001dd27779a009e580066cf5aafb24a665bc37"} Mar 13 08:00:03 crc kubenswrapper[4876]: I0313 08:00:03.234786 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 08:00:03 crc kubenswrapper[4876]: I0313 08:00:03.334462 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 08:00:03 crc kubenswrapper[4876]: I0313 08:00:03.346515 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6dbd6d4d74-7zn9n" Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.639014 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.757359 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjl8s\" (UniqueName: \"kubernetes.io/projected/556fa6a5-90ac-43ea-a365-abd7ea03f21a-kube-api-access-pjl8s\") pod \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.757462 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556fa6a5-90ac-43ea-a365-abd7ea03f21a-config-volume\") pod \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.757512 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556fa6a5-90ac-43ea-a365-abd7ea03f21a-secret-volume\") pod \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\" (UID: \"556fa6a5-90ac-43ea-a365-abd7ea03f21a\") " Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.761255 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/556fa6a5-90ac-43ea-a365-abd7ea03f21a-config-volume" (OuterVolumeSpecName: "config-volume") pod "556fa6a5-90ac-43ea-a365-abd7ea03f21a" (UID: "556fa6a5-90ac-43ea-a365-abd7ea03f21a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.770451 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/556fa6a5-90ac-43ea-a365-abd7ea03f21a-kube-api-access-pjl8s" (OuterVolumeSpecName: "kube-api-access-pjl8s") pod "556fa6a5-90ac-43ea-a365-abd7ea03f21a" (UID: "556fa6a5-90ac-43ea-a365-abd7ea03f21a"). InnerVolumeSpecName "kube-api-access-pjl8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.805567 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556fa6a5-90ac-43ea-a365-abd7ea03f21a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "556fa6a5-90ac-43ea-a365-abd7ea03f21a" (UID: "556fa6a5-90ac-43ea-a365-abd7ea03f21a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.860796 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjl8s\" (UniqueName: \"kubernetes.io/projected/556fa6a5-90ac-43ea-a365-abd7ea03f21a-kube-api-access-pjl8s\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.860837 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/556fa6a5-90ac-43ea-a365-abd7ea03f21a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:04 crc kubenswrapper[4876]: I0313 08:00:04.860848 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/556fa6a5-90ac-43ea-a365-abd7ea03f21a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.159990 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" event={"ID":"556fa6a5-90ac-43ea-a365-abd7ea03f21a","Type":"ContainerDied","Data":"fc7bb76035c345d0e092b1f79701c4c42304c805e7a77fc98167b3dfe6b720c3"} Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.160549 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc7bb76035c345d0e092b1f79701c4c42304c805e7a77fc98167b3dfe6b720c3" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.160616 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.198773 4876 generic.go:334] "Generic (PLEG): container finished" podID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerID="fe205b2df6edfb12f85b95b867ce4fc59c121522b1f5725082c986daf4eb8f74" exitCode=0 Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.198857 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-858fd4b7b-stb59" event={"ID":"4a2479e9-5256-4e71-8f79-44cebfb76fa7","Type":"ContainerDied","Data":"fe205b2df6edfb12f85b95b867ce4fc59c121522b1f5725082c986daf4eb8f74"} Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.431373 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.537089 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4vjj\" (UniqueName: \"kubernetes.io/projected/4a2479e9-5256-4e71-8f79-44cebfb76fa7-kube-api-access-w4vjj\") pod \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.537212 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a2479e9-5256-4e71-8f79-44cebfb76fa7-logs\") pod \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.537411 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data\") pod \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.537507 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data-custom\") pod \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.537619 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-combined-ca-bundle\") pod \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\" (UID: \"4a2479e9-5256-4e71-8f79-44cebfb76fa7\") " Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.549522 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a2479e9-5256-4e71-8f79-44cebfb76fa7-kube-api-access-w4vjj" (OuterVolumeSpecName: "kube-api-access-w4vjj") pod "4a2479e9-5256-4e71-8f79-44cebfb76fa7" (UID: "4a2479e9-5256-4e71-8f79-44cebfb76fa7"). InnerVolumeSpecName "kube-api-access-w4vjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.562755 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4a2479e9-5256-4e71-8f79-44cebfb76fa7" (UID: "4a2479e9-5256-4e71-8f79-44cebfb76fa7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.570541 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a2479e9-5256-4e71-8f79-44cebfb76fa7-logs" (OuterVolumeSpecName: "logs") pod "4a2479e9-5256-4e71-8f79-44cebfb76fa7" (UID: "4a2479e9-5256-4e71-8f79-44cebfb76fa7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.639532 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a2479e9-5256-4e71-8f79-44cebfb76fa7" (UID: "4a2479e9-5256-4e71-8f79-44cebfb76fa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.642267 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.642293 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.642307 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4vjj\" (UniqueName: \"kubernetes.io/projected/4a2479e9-5256-4e71-8f79-44cebfb76fa7-kube-api-access-w4vjj\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.642318 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a2479e9-5256-4e71-8f79-44cebfb76fa7-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.727429 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data" (OuterVolumeSpecName: "config-data") pod "4a2479e9-5256-4e71-8f79-44cebfb76fa7" (UID: "4a2479e9-5256-4e71-8f79-44cebfb76fa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:05 crc kubenswrapper[4876]: I0313 08:00:05.750077 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a2479e9-5256-4e71-8f79-44cebfb76fa7-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.077306 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7b9c5d6d4c-7lxwx"] Mar 13 08:00:06 crc kubenswrapper[4876]: E0313 08:00:06.077776 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerName="barbican-api-log" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.077795 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerName="barbican-api-log" Mar 13 08:00:06 crc kubenswrapper[4876]: E0313 08:00:06.077828 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerName="barbican-api" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.077836 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerName="barbican-api" Mar 13 08:00:06 crc kubenswrapper[4876]: E0313 08:00:06.077845 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="556fa6a5-90ac-43ea-a365-abd7ea03f21a" containerName="collect-profiles" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.077851 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="556fa6a5-90ac-43ea-a365-abd7ea03f21a" containerName="collect-profiles" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.078056 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="556fa6a5-90ac-43ea-a365-abd7ea03f21a" containerName="collect-profiles" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.078074 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerName="barbican-api-log" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.078093 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" containerName="barbican-api" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.079331 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.081693 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.081875 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.082477 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.151545 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b9c5d6d4c-7lxwx"] Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.160959 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-log-httpd\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.161112 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-etc-swift\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.161183 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-combined-ca-bundle\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.161211 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-run-httpd\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.161275 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-config-data\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.161321 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-public-tls-certs\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.161356 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-internal-tls-certs\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.161399 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gltgw\" (UniqueName: \"kubernetes.io/projected/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-kube-api-access-gltgw\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.215604 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-858fd4b7b-stb59" event={"ID":"4a2479e9-5256-4e71-8f79-44cebfb76fa7","Type":"ContainerDied","Data":"f97161a4ebe6b77286166276d77ab64a2abee0cb4e24e1191be8df52c6cbb398"} Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.215685 4876 scope.go:117] "RemoveContainer" containerID="fe205b2df6edfb12f85b95b867ce4fc59c121522b1f5725082c986daf4eb8f74" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.215853 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-858fd4b7b-stb59" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.263335 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-etc-swift\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.263432 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-combined-ca-bundle\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.263467 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-run-httpd\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.263504 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-config-data\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.263556 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-public-tls-certs\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.263589 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-internal-tls-certs\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.263632 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gltgw\" (UniqueName: \"kubernetes.io/projected/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-kube-api-access-gltgw\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.263694 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-log-httpd\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.264269 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-log-httpd\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.265406 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-858fd4b7b-stb59"] Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.268025 4876 scope.go:117] "RemoveContainer" containerID="df53d39702dc0583b7cefe144f4f08804c1dfa962ef426cd072075b277e9921e" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.269173 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-run-httpd\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.270074 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-etc-swift\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.270642 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-config-data\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.271168 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-internal-tls-certs\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.271831 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-combined-ca-bundle\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.277192 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-public-tls-certs\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.279751 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-858fd4b7b-stb59"] Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.286004 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gltgw\" (UniqueName: \"kubernetes.io/projected/62fd9eb3-6af0-4bc9-bc57-65c8b03eb467-kube-api-access-gltgw\") pod \"swift-proxy-7b9c5d6d4c-7lxwx\" (UID: \"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467\") " pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.406950 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.885156 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.885888 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="ceilometer-central-agent" containerID="cri-o://5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621" gracePeriod=30 Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.886044 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="sg-core" containerID="cri-o://0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3" gracePeriod=30 Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.886044 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="proxy-httpd" containerID="cri-o://73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec" gracePeriod=30 Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.886069 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="ceilometer-notification-agent" containerID="cri-o://e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553" gracePeriod=30 Mar 13 08:00:06 crc kubenswrapper[4876]: I0313 08:00:06.893896 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.046610 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a2479e9-5256-4e71-8f79-44cebfb76fa7" path="/var/lib/kubelet/pods/4a2479e9-5256-4e71-8f79-44cebfb76fa7/volumes" Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.112230 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b9c5d6d4c-7lxwx"] Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.246601 4876 generic.go:334] "Generic (PLEG): container finished" podID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerID="73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec" exitCode=0 Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.246659 4876 generic.go:334] "Generic (PLEG): container finished" podID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerID="0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3" exitCode=2 Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.246697 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerDied","Data":"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec"} Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.246776 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerDied","Data":"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3"} Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.250912 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" event={"ID":"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467","Type":"ContainerStarted","Data":"6177014f2561786003c42a49bd1cf27491f013e72e2b45a857f8bceca59b1212"} Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.857047 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.913477 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-run-httpd\") pod \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.913561 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-log-httpd\") pod \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.913745 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-config-data\") pod \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.913788 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-sg-core-conf-yaml\") pod \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.913879 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-combined-ca-bundle\") pod \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.913934 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kmlp\" (UniqueName: \"kubernetes.io/projected/f43b9345-cbf5-4885-9ac4-8e944d8d3142-kube-api-access-2kmlp\") pod \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.913998 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-scripts\") pod \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\" (UID: \"f43b9345-cbf5-4885-9ac4-8e944d8d3142\") " Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.914539 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f43b9345-cbf5-4885-9ac4-8e944d8d3142" (UID: "f43b9345-cbf5-4885-9ac4-8e944d8d3142"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.915697 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f43b9345-cbf5-4885-9ac4-8e944d8d3142" (UID: "f43b9345-cbf5-4885-9ac4-8e944d8d3142"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.923911 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-scripts" (OuterVolumeSpecName: "scripts") pod "f43b9345-cbf5-4885-9ac4-8e944d8d3142" (UID: "f43b9345-cbf5-4885-9ac4-8e944d8d3142"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.925110 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43b9345-cbf5-4885-9ac4-8e944d8d3142-kube-api-access-2kmlp" (OuterVolumeSpecName: "kube-api-access-2kmlp") pod "f43b9345-cbf5-4885-9ac4-8e944d8d3142" (UID: "f43b9345-cbf5-4885-9ac4-8e944d8d3142"). InnerVolumeSpecName "kube-api-access-2kmlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:07 crc kubenswrapper[4876]: I0313 08:00:07.958727 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f43b9345-cbf5-4885-9ac4-8e944d8d3142" (UID: "f43b9345-cbf5-4885-9ac4-8e944d8d3142"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.022954 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.022995 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.023008 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kmlp\" (UniqueName: \"kubernetes.io/projected/f43b9345-cbf5-4885-9ac4-8e944d8d3142-kube-api-access-2kmlp\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.023028 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.023040 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f43b9345-cbf5-4885-9ac4-8e944d8d3142-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.035317 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f43b9345-cbf5-4885-9ac4-8e944d8d3142" (UID: "f43b9345-cbf5-4885-9ac4-8e944d8d3142"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.071873 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-config-data" (OuterVolumeSpecName: "config-data") pod "f43b9345-cbf5-4885-9ac4-8e944d8d3142" (UID: "f43b9345-cbf5-4885-9ac4-8e944d8d3142"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.125364 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.125396 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43b9345-cbf5-4885-9ac4-8e944d8d3142-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.273955 4876 generic.go:334] "Generic (PLEG): container finished" podID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerID="e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553" exitCode=0 Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.273996 4876 generic.go:334] "Generic (PLEG): container finished" podID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerID="5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621" exitCode=0 Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.274047 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerDied","Data":"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553"} Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.274084 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerDied","Data":"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621"} Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.274096 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f43b9345-cbf5-4885-9ac4-8e944d8d3142","Type":"ContainerDied","Data":"87ada5e68ea0261af0a10cb1d652a149eaead4ad156ccf9bea4c23ae75fa8b7a"} Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.274116 4876 scope.go:117] "RemoveContainer" containerID="73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.274266 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.283912 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" event={"ID":"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467","Type":"ContainerStarted","Data":"fd922bbe5e6ead32b37048d01aebe834975b3ff6c92e2eb52a6f03eca217bd6a"} Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.283973 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" event={"ID":"62fd9eb3-6af0-4bc9-bc57-65c8b03eb467","Type":"ContainerStarted","Data":"5879faa66cbf14d43e48a454d1fc9df6918153787e253ebe6e7e7995484c84c8"} Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.284282 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.351511 4876 scope.go:117] "RemoveContainer" containerID="0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.392422 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" podStartSLOduration=2.392397012 podStartE2EDuration="2.392397012s" podCreationTimestamp="2026-03-13 08:00:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:08.357437805 +0000 UTC m=+1268.028216807" watchObservedRunningTime="2026-03-13 08:00:08.392397012 +0000 UTC m=+1268.063175994" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.400722 4876 scope.go:117] "RemoveContainer" containerID="e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.414598 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.426732 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.465034 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:08 crc kubenswrapper[4876]: E0313 08:00:08.465917 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="sg-core" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.465982 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="sg-core" Mar 13 08:00:08 crc kubenswrapper[4876]: E0313 08:00:08.466045 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="ceilometer-notification-agent" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.466054 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="ceilometer-notification-agent" Mar 13 08:00:08 crc kubenswrapper[4876]: E0313 08:00:08.466063 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="ceilometer-central-agent" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.466072 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="ceilometer-central-agent" Mar 13 08:00:08 crc kubenswrapper[4876]: E0313 08:00:08.466087 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="proxy-httpd" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.466095 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="proxy-httpd" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.466303 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="ceilometer-notification-agent" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.466338 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="sg-core" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.466538 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="ceilometer-central-agent" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.466552 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" containerName="proxy-httpd" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.469170 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.477320 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.477320 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.483560 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.506371 4876 scope.go:117] "RemoveContainer" containerID="5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.530950 4876 scope.go:117] "RemoveContainer" containerID="73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec" Mar 13 08:00:08 crc kubenswrapper[4876]: E0313 08:00:08.533489 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec\": container with ID starting with 73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec not found: ID does not exist" containerID="73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.533644 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec"} err="failed to get container status \"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec\": rpc error: code = NotFound desc = could not find container \"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec\": container with ID starting with 73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec not found: ID does not exist" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.533682 4876 scope.go:117] "RemoveContainer" containerID="0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3" Mar 13 08:00:08 crc kubenswrapper[4876]: E0313 08:00:08.534168 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3\": container with ID starting with 0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3 not found: ID does not exist" containerID="0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.534191 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3"} err="failed to get container status \"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3\": rpc error: code = NotFound desc = could not find container \"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3\": container with ID starting with 0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3 not found: ID does not exist" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.534207 4876 scope.go:117] "RemoveContainer" containerID="e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553" Mar 13 08:00:08 crc kubenswrapper[4876]: E0313 08:00:08.534715 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553\": container with ID starting with e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553 not found: ID does not exist" containerID="e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.534740 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553"} err="failed to get container status \"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553\": rpc error: code = NotFound desc = could not find container \"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553\": container with ID starting with e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553 not found: ID does not exist" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.534757 4876 scope.go:117] "RemoveContainer" containerID="5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621" Mar 13 08:00:08 crc kubenswrapper[4876]: E0313 08:00:08.535054 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621\": container with ID starting with 5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621 not found: ID does not exist" containerID="5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.535085 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621"} err="failed to get container status \"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621\": rpc error: code = NotFound desc = could not find container \"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621\": container with ID starting with 5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621 not found: ID does not exist" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.535102 4876 scope.go:117] "RemoveContainer" containerID="73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.535642 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec"} err="failed to get container status \"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec\": rpc error: code = NotFound desc = could not find container \"73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec\": container with ID starting with 73f7949462983f7f8a46375dae3c85503b8411b01497623b0ee860a69d152aec not found: ID does not exist" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.535666 4876 scope.go:117] "RemoveContainer" containerID="0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.536230 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3"} err="failed to get container status \"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3\": rpc error: code = NotFound desc = could not find container \"0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3\": container with ID starting with 0df935bb63ca958d025365a5b9a9e6d49c7257d4a8666ceebebbb28f8e7399d3 not found: ID does not exist" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.536284 4876 scope.go:117] "RemoveContainer" containerID="e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.536736 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553"} err="failed to get container status \"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553\": rpc error: code = NotFound desc = could not find container \"e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553\": container with ID starting with e49a41e810d68b610e189337ed904866ff68991cb54463796eaa17b68481f553 not found: ID does not exist" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.536791 4876 scope.go:117] "RemoveContainer" containerID="5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.537189 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621"} err="failed to get container status \"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621\": rpc error: code = NotFound desc = could not find container \"5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621\": container with ID starting with 5770161d2a946373eda6ab16fb033ef3aa020fcc696d80f75ba64b1cae7bb621 not found: ID does not exist" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.552498 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.552583 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-run-httpd\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.552638 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.552688 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-config-data\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.552771 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz6hk\" (UniqueName: \"kubernetes.io/projected/018ec9da-cec8-4c3c-a5b7-c652ef88de04-kube-api-access-jz6hk\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.552819 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-log-httpd\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.552847 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-scripts\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.584632 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.655526 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-config-data\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.655666 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz6hk\" (UniqueName: \"kubernetes.io/projected/018ec9da-cec8-4c3c-a5b7-c652ef88de04-kube-api-access-jz6hk\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.655721 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-log-httpd\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.655760 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-scripts\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.655804 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.655874 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-run-httpd\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.655923 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.657399 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-log-httpd\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.657440 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-run-httpd\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.660162 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.661127 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-scripts\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.661532 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.665158 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-config-data\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.679411 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz6hk\" (UniqueName: \"kubernetes.io/projected/018ec9da-cec8-4c3c-a5b7-c652ef88de04-kube-api-access-jz6hk\") pod \"ceilometer-0\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " pod="openstack/ceilometer-0" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.718150 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66dd85954b-6789x" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 08:00:08 crc kubenswrapper[4876]: I0313 08:00:08.811051 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:00:09 crc kubenswrapper[4876]: I0313 08:00:09.061351 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43b9345-cbf5-4885-9ac4-8e944d8d3142" path="/var/lib/kubelet/pods/f43b9345-cbf5-4885-9ac4-8e944d8d3142/volumes" Mar 13 08:00:09 crc kubenswrapper[4876]: I0313 08:00:09.325343 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:09 crc kubenswrapper[4876]: I0313 08:00:09.376431 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:10 crc kubenswrapper[4876]: I0313 08:00:10.681283 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 08:00:10 crc kubenswrapper[4876]: I0313 08:00:10.682287 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="17347053-1f92-46e8-a1e2-be471cf19516" containerName="glance-log" containerID="cri-o://67d5719f159209f2dad8684db9e7b94b6e9dc28cd3cc9f5d8b5c4ddf4962f041" gracePeriod=30 Mar 13 08:00:10 crc kubenswrapper[4876]: I0313 08:00:10.682772 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="17347053-1f92-46e8-a1e2-be471cf19516" containerName="glance-httpd" containerID="cri-o://9d9755e38fd10bcc26796d4ab4445627fb53701c36a6fd0b6a7e471def88f7e9" gracePeriod=30 Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.215492 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5884ccb8d9-4wjtm" Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.313813 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5cb9cf56c6-k7vkg"] Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.314373 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5cb9cf56c6-k7vkg" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerName="neutron-api" containerID="cri-o://78a12f079c86c03a1434cab042a6e40ad549af96a3e8a139ac0f74fb201bedce" gracePeriod=30 Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.314995 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5cb9cf56c6-k7vkg" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerName="neutron-httpd" containerID="cri-o://50e23ad80caa9b0a975b0d31c0f0ca3c2d7d59e5b837f13c9b725b68dbce8a4e" gracePeriod=30 Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.370346 4876 generic.go:334] "Generic (PLEG): container finished" podID="17347053-1f92-46e8-a1e2-be471cf19516" containerID="67d5719f159209f2dad8684db9e7b94b6e9dc28cd3cc9f5d8b5c4ddf4962f041" exitCode=143 Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.370441 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17347053-1f92-46e8-a1e2-be471cf19516","Type":"ContainerDied","Data":"67d5719f159209f2dad8684db9e7b94b6e9dc28cd3cc9f5d8b5c4ddf4962f041"} Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.961632 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.963769 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b3664dea-7307-496d-aff3-588911d155bb" containerName="glance-httpd" containerID="cri-o://ac6988d9ae894d1f9bb1fe44270e4e0f5f2765ea43b1cc4650101b6c16699110" gracePeriod=30 Mar 13 08:00:11 crc kubenswrapper[4876]: I0313 08:00:11.968672 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b3664dea-7307-496d-aff3-588911d155bb" containerName="glance-log" containerID="cri-o://8862cfa4ed45d9337b0f1fd8dc684f0790bbcc8d758836f140c02da1ae2631ea" gracePeriod=30 Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.016466 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-x9tx5"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.017753 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.041519 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x9tx5"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.136101 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-kl96z"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.137705 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.154210 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-102d-account-create-update-kvlpr"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.155906 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.157183 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhjgz\" (UniqueName: \"kubernetes.io/projected/51b8e947-a71d-4365-a028-7f8d6d184577-kube-api-access-zhjgz\") pod \"nova-api-db-create-x9tx5\" (UID: \"51b8e947-a71d-4365-a028-7f8d6d184577\") " pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.157295 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51b8e947-a71d-4365-a028-7f8d6d184577-operator-scripts\") pod \"nova-api-db-create-x9tx5\" (UID: \"51b8e947-a71d-4365-a028-7f8d6d184577\") " pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.160273 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.164511 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-kl96z"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.174988 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-102d-account-create-update-kvlpr"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.260142 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tn74\" (UniqueName: \"kubernetes.io/projected/6e0939f9-b421-4707-861c-b43fae99f079-kube-api-access-6tn74\") pod \"nova-cell0-db-create-kl96z\" (UID: \"6e0939f9-b421-4707-861c-b43fae99f079\") " pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.260192 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0939f9-b421-4707-861c-b43fae99f079-operator-scripts\") pod \"nova-cell0-db-create-kl96z\" (UID: \"6e0939f9-b421-4707-861c-b43fae99f079\") " pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.260275 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhjgz\" (UniqueName: \"kubernetes.io/projected/51b8e947-a71d-4365-a028-7f8d6d184577-kube-api-access-zhjgz\") pod \"nova-api-db-create-x9tx5\" (UID: \"51b8e947-a71d-4365-a028-7f8d6d184577\") " pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.260696 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51b8e947-a71d-4365-a028-7f8d6d184577-operator-scripts\") pod \"nova-api-db-create-x9tx5\" (UID: \"51b8e947-a71d-4365-a028-7f8d6d184577\") " pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.260758 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvsrr\" (UniqueName: \"kubernetes.io/projected/3fec4b3c-6d95-4ed5-ac57-40da620b201c-kube-api-access-rvsrr\") pod \"nova-api-102d-account-create-update-kvlpr\" (UID: \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\") " pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.260980 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fec4b3c-6d95-4ed5-ac57-40da620b201c-operator-scripts\") pod \"nova-api-102d-account-create-update-kvlpr\" (UID: \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\") " pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.262431 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51b8e947-a71d-4365-a028-7f8d6d184577-operator-scripts\") pod \"nova-api-db-create-x9tx5\" (UID: \"51b8e947-a71d-4365-a028-7f8d6d184577\") " pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.293019 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhjgz\" (UniqueName: \"kubernetes.io/projected/51b8e947-a71d-4365-a028-7f8d6d184577-kube-api-access-zhjgz\") pod \"nova-api-db-create-x9tx5\" (UID: \"51b8e947-a71d-4365-a028-7f8d6d184577\") " pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.338909 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-wxvxg"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.340431 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.359028 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-3c87-account-create-update-mgw9g"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.366546 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.367581 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvsrr\" (UniqueName: \"kubernetes.io/projected/3fec4b3c-6d95-4ed5-ac57-40da620b201c-kube-api-access-rvsrr\") pod \"nova-api-102d-account-create-update-kvlpr\" (UID: \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\") " pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.367759 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fec4b3c-6d95-4ed5-ac57-40da620b201c-operator-scripts\") pod \"nova-api-102d-account-create-update-kvlpr\" (UID: \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\") " pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.367970 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tn74\" (UniqueName: \"kubernetes.io/projected/6e0939f9-b421-4707-861c-b43fae99f079-kube-api-access-6tn74\") pod \"nova-cell0-db-create-kl96z\" (UID: \"6e0939f9-b421-4707-861c-b43fae99f079\") " pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.367997 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0939f9-b421-4707-861c-b43fae99f079-operator-scripts\") pod \"nova-cell0-db-create-kl96z\" (UID: \"6e0939f9-b421-4707-861c-b43fae99f079\") " pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.369003 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0939f9-b421-4707-861c-b43fae99f079-operator-scripts\") pod \"nova-cell0-db-create-kl96z\" (UID: \"6e0939f9-b421-4707-861c-b43fae99f079\") " pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.369576 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fec4b3c-6d95-4ed5-ac57-40da620b201c-operator-scripts\") pod \"nova-api-102d-account-create-update-kvlpr\" (UID: \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\") " pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.370959 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.372552 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wxvxg"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.383677 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3c87-account-create-update-mgw9g"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.402772 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.453966 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tn74\" (UniqueName: \"kubernetes.io/projected/6e0939f9-b421-4707-861c-b43fae99f079-kube-api-access-6tn74\") pod \"nova-cell0-db-create-kl96z\" (UID: \"6e0939f9-b421-4707-861c-b43fae99f079\") " pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.454507 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvsrr\" (UniqueName: \"kubernetes.io/projected/3fec4b3c-6d95-4ed5-ac57-40da620b201c-kube-api-access-rvsrr\") pod \"nova-api-102d-account-create-update-kvlpr\" (UID: \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\") " pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.474619 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6kgp\" (UniqueName: \"kubernetes.io/projected/8229f43e-63f9-41a0-8197-4af96c27ee48-kube-api-access-w6kgp\") pod \"nova-cell0-3c87-account-create-update-mgw9g\" (UID: \"8229f43e-63f9-41a0-8197-4af96c27ee48\") " pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.474721 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmqgd\" (UniqueName: \"kubernetes.io/projected/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-kube-api-access-nmqgd\") pod \"nova-cell1-db-create-wxvxg\" (UID: \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\") " pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.474813 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-operator-scripts\") pod \"nova-cell1-db-create-wxvxg\" (UID: \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\") " pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.475300 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8229f43e-63f9-41a0-8197-4af96c27ee48-operator-scripts\") pod \"nova-cell0-3c87-account-create-update-mgw9g\" (UID: \"8229f43e-63f9-41a0-8197-4af96c27ee48\") " pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.476417 4876 generic.go:334] "Generic (PLEG): container finished" podID="b3664dea-7307-496d-aff3-588911d155bb" containerID="8862cfa4ed45d9337b0f1fd8dc684f0790bbcc8d758836f140c02da1ae2631ea" exitCode=143 Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.476597 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3664dea-7307-496d-aff3-588911d155bb","Type":"ContainerDied","Data":"8862cfa4ed45d9337b0f1fd8dc684f0790bbcc8d758836f140c02da1ae2631ea"} Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.479712 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.485778 4876 generic.go:334] "Generic (PLEG): container finished" podID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerID="50e23ad80caa9b0a975b0d31c0f0ca3c2d7d59e5b837f13c9b725b68dbce8a4e" exitCode=0 Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.485824 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cb9cf56c6-k7vkg" event={"ID":"9fd26454-5f72-4e52-a3e7-87270228d46d","Type":"ContainerDied","Data":"50e23ad80caa9b0a975b0d31c0f0ca3c2d7d59e5b837f13c9b725b68dbce8a4e"} Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.494229 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.527844 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-75fc-account-create-update-zjt7n"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.529461 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.533916 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.536064 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-75fc-account-create-update-zjt7n"] Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.577744 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-operator-scripts\") pod \"nova-cell1-db-create-wxvxg\" (UID: \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\") " pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.577915 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8229f43e-63f9-41a0-8197-4af96c27ee48-operator-scripts\") pod \"nova-cell0-3c87-account-create-update-mgw9g\" (UID: \"8229f43e-63f9-41a0-8197-4af96c27ee48\") " pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.577972 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6kgp\" (UniqueName: \"kubernetes.io/projected/8229f43e-63f9-41a0-8197-4af96c27ee48-kube-api-access-w6kgp\") pod \"nova-cell0-3c87-account-create-update-mgw9g\" (UID: \"8229f43e-63f9-41a0-8197-4af96c27ee48\") " pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.577999 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmqgd\" (UniqueName: \"kubernetes.io/projected/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-kube-api-access-nmqgd\") pod \"nova-cell1-db-create-wxvxg\" (UID: \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\") " pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.579263 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8229f43e-63f9-41a0-8197-4af96c27ee48-operator-scripts\") pod \"nova-cell0-3c87-account-create-update-mgw9g\" (UID: \"8229f43e-63f9-41a0-8197-4af96c27ee48\") " pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.579835 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-operator-scripts\") pod \"nova-cell1-db-create-wxvxg\" (UID: \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\") " pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.610935 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6kgp\" (UniqueName: \"kubernetes.io/projected/8229f43e-63f9-41a0-8197-4af96c27ee48-kube-api-access-w6kgp\") pod \"nova-cell0-3c87-account-create-update-mgw9g\" (UID: \"8229f43e-63f9-41a0-8197-4af96c27ee48\") " pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.632567 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmqgd\" (UniqueName: \"kubernetes.io/projected/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-kube-api-access-nmqgd\") pod \"nova-cell1-db-create-wxvxg\" (UID: \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\") " pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.683349 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-operator-scripts\") pod \"nova-cell1-75fc-account-create-update-zjt7n\" (UID: \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\") " pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.683523 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jv9w\" (UniqueName: \"kubernetes.io/projected/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-kube-api-access-2jv9w\") pod \"nova-cell1-75fc-account-create-update-zjt7n\" (UID: \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\") " pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.688829 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.779505 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.785640 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jv9w\" (UniqueName: \"kubernetes.io/projected/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-kube-api-access-2jv9w\") pod \"nova-cell1-75fc-account-create-update-zjt7n\" (UID: \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\") " pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.785817 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-operator-scripts\") pod \"nova-cell1-75fc-account-create-update-zjt7n\" (UID: \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\") " pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.786470 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-operator-scripts\") pod \"nova-cell1-75fc-account-create-update-zjt7n\" (UID: \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\") " pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.810187 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jv9w\" (UniqueName: \"kubernetes.io/projected/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-kube-api-access-2jv9w\") pod \"nova-cell1-75fc-account-create-update-zjt7n\" (UID: \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\") " pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:12 crc kubenswrapper[4876]: I0313 08:00:12.849808 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:13 crc kubenswrapper[4876]: I0313 08:00:13.295616 4876 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod4e7e7ea3-8a44-41f4-a411-fa8c5abf1d5a] : Timed out while waiting for systemd to remove kubepods-besteffort-pod4e7e7ea3_8a44_41f4_a411_fa8c5abf1d5a.slice" Mar 13 08:00:13 crc kubenswrapper[4876]: I0313 08:00:13.821134 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:14 crc kubenswrapper[4876]: I0313 08:00:14.510394 4876 generic.go:334] "Generic (PLEG): container finished" podID="17347053-1f92-46e8-a1e2-be471cf19516" containerID="9d9755e38fd10bcc26796d4ab4445627fb53701c36a6fd0b6a7e471def88f7e9" exitCode=0 Mar 13 08:00:14 crc kubenswrapper[4876]: I0313 08:00:14.510475 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17347053-1f92-46e8-a1e2-be471cf19516","Type":"ContainerDied","Data":"9d9755e38fd10bcc26796d4ab4445627fb53701c36a6fd0b6a7e471def88f7e9"} Mar 13 08:00:15 crc kubenswrapper[4876]: I0313 08:00:15.525566 4876 generic.go:334] "Generic (PLEG): container finished" podID="b3664dea-7307-496d-aff3-588911d155bb" containerID="ac6988d9ae894d1f9bb1fe44270e4e0f5f2765ea43b1cc4650101b6c16699110" exitCode=0 Mar 13 08:00:15 crc kubenswrapper[4876]: I0313 08:00:15.525656 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3664dea-7307-496d-aff3-588911d155bb","Type":"ContainerDied","Data":"ac6988d9ae894d1f9bb1fe44270e4e0f5f2765ea43b1cc4650101b6c16699110"} Mar 13 08:00:16 crc kubenswrapper[4876]: I0313 08:00:16.414330 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:16 crc kubenswrapper[4876]: I0313 08:00:16.416088 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" Mar 13 08:00:17 crc kubenswrapper[4876]: E0313 08:00:17.103566 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c0e0e38_68d2_45a0_9601_fe2e4f3f9179.slice/crio-conmon-f759362668103ca973b7087882f219d00529c8d9e78138c3c9d8b0b0451038b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3664dea_7307_496d_aff3_588911d155bb.slice/crio-ac6988d9ae894d1f9bb1fe44270e4e0f5f2765ea43b1cc4650101b6c16699110.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17347053_1f92_46e8_a1e2_be471cf19516.slice/crio-conmon-9d9755e38fd10bcc26796d4ab4445627fb53701c36a6fd0b6a7e471def88f7e9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c0e0e38_68d2_45a0_9601_fe2e4f3f9179.slice/crio-f759362668103ca973b7087882f219d00529c8d9e78138c3c9d8b0b0451038b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3664dea_7307_496d_aff3_588911d155bb.slice/crio-conmon-ac6988d9ae894d1f9bb1fe44270e4e0f5f2765ea43b1cc4650101b6c16699110.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17347053_1f92_46e8_a1e2_be471cf19516.slice/crio-9d9755e38fd10bcc26796d4ab4445627fb53701c36a6fd0b6a7e471def88f7e9.scope\": RecentStats: unable to find data in memory cache]" Mar 13 08:00:17 crc kubenswrapper[4876]: W0313 08:00:17.495374 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod018ec9da_cec8_4c3c_a5b7_c652ef88de04.slice/crio-45afb5de0caf327d6aea6bb3637e47878a0d5667121b8a79ff5ee3a3f6da7c85 WatchSource:0}: Error finding container 45afb5de0caf327d6aea6bb3637e47878a0d5667121b8a79ff5ee3a3f6da7c85: Status 404 returned error can't find the container with id 45afb5de0caf327d6aea6bb3637e47878a0d5667121b8a79ff5ee3a3f6da7c85 Mar 13 08:00:17 crc kubenswrapper[4876]: I0313 08:00:17.609296 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerStarted","Data":"45afb5de0caf327d6aea6bb3637e47878a0d5667121b8a79ff5ee3a3f6da7c85"} Mar 13 08:00:17 crc kubenswrapper[4876]: I0313 08:00:17.615694 4876 generic.go:334] "Generic (PLEG): container finished" podID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerID="f759362668103ca973b7087882f219d00529c8d9e78138c3c9d8b0b0451038b6" exitCode=137 Mar 13 08:00:17 crc kubenswrapper[4876]: I0313 08:00:17.615835 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179","Type":"ContainerDied","Data":"f759362668103ca973b7087882f219d00529c8d9e78138c3c9d8b0b0451038b6"} Mar 13 08:00:17 crc kubenswrapper[4876]: I0313 08:00:17.623303 4876 generic.go:334] "Generic (PLEG): container finished" podID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerID="78a12f079c86c03a1434cab042a6e40ad549af96a3e8a139ac0f74fb201bedce" exitCode=0 Mar 13 08:00:17 crc kubenswrapper[4876]: I0313 08:00:17.623368 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cb9cf56c6-k7vkg" event={"ID":"9fd26454-5f72-4e52-a3e7-87270228d46d","Type":"ContainerDied","Data":"78a12f079c86c03a1434cab042a6e40ad549af96a3e8a139ac0f74fb201bedce"} Mar 13 08:00:17 crc kubenswrapper[4876]: I0313 08:00:17.910387 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.051080 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr4jt\" (UniqueName: \"kubernetes.io/projected/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-kube-api-access-mr4jt\") pod \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.051123 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data\") pod \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.051176 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-logs\") pod \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.051228 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data-custom\") pod \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.051277 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-combined-ca-bundle\") pod \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.051346 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-scripts\") pod \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.051500 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-etc-machine-id\") pod \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\" (UID: \"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.052010 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" (UID: "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.054927 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-logs" (OuterVolumeSpecName: "logs") pod "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" (UID: "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.068160 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" (UID: "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.079554 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-kube-api-access-mr4jt" (OuterVolumeSpecName: "kube-api-access-mr4jt") pod "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" (UID: "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179"). InnerVolumeSpecName "kube-api-access-mr4jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.088098 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-scripts" (OuterVolumeSpecName: "scripts") pod "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" (UID: "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.138357 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" (UID: "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.153743 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr4jt\" (UniqueName: \"kubernetes.io/projected/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-kube-api-access-mr4jt\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.153794 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.153808 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.153816 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.153826 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.153835 4876 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.177568 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data" (OuterVolumeSpecName: "config-data") pod "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" (UID: "6c0e0e38-68d2-45a0-9601-fe2e4f3f9179"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.255488 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.368264 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.459573 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x9tx5"] Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.460137 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-combined-ca-bundle\") pod \"17347053-1f92-46e8-a1e2-be471cf19516\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.460295 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-scripts\") pod \"17347053-1f92-46e8-a1e2-be471cf19516\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.460331 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk264\" (UniqueName: \"kubernetes.io/projected/17347053-1f92-46e8-a1e2-be471cf19516-kube-api-access-dk264\") pod \"17347053-1f92-46e8-a1e2-be471cf19516\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.460391 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-httpd-run\") pod \"17347053-1f92-46e8-a1e2-be471cf19516\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.460425 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"17347053-1f92-46e8-a1e2-be471cf19516\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.460513 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-public-tls-certs\") pod \"17347053-1f92-46e8-a1e2-be471cf19516\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.460609 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-logs\") pod \"17347053-1f92-46e8-a1e2-be471cf19516\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.460635 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-config-data\") pod \"17347053-1f92-46e8-a1e2-be471cf19516\" (UID: \"17347053-1f92-46e8-a1e2-be471cf19516\") " Mar 13 08:00:18 crc kubenswrapper[4876]: W0313 08:00:18.460635 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51b8e947_a71d_4365_a028_7f8d6d184577.slice/crio-c8cc86a038669528c46ce81fd244531f11ea9f8df5e61a2c7d8c751e9036f9a4 WatchSource:0}: Error finding container c8cc86a038669528c46ce81fd244531f11ea9f8df5e61a2c7d8c751e9036f9a4: Status 404 returned error can't find the container with id c8cc86a038669528c46ce81fd244531f11ea9f8df5e61a2c7d8c751e9036f9a4 Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.463740 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "17347053-1f92-46e8-a1e2-be471cf19516" (UID: "17347053-1f92-46e8-a1e2-be471cf19516"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.466833 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-logs" (OuterVolumeSpecName: "logs") pod "17347053-1f92-46e8-a1e2-be471cf19516" (UID: "17347053-1f92-46e8-a1e2-be471cf19516"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.469103 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17347053-1f92-46e8-a1e2-be471cf19516-kube-api-access-dk264" (OuterVolumeSpecName: "kube-api-access-dk264") pod "17347053-1f92-46e8-a1e2-be471cf19516" (UID: "17347053-1f92-46e8-a1e2-be471cf19516"). InnerVolumeSpecName "kube-api-access-dk264". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.479858 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-scripts" (OuterVolumeSpecName: "scripts") pod "17347053-1f92-46e8-a1e2-be471cf19516" (UID: "17347053-1f92-46e8-a1e2-be471cf19516"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.492501 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "17347053-1f92-46e8-a1e2-be471cf19516" (UID: "17347053-1f92-46e8-a1e2-be471cf19516"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.555372 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17347053-1f92-46e8-a1e2-be471cf19516" (UID: "17347053-1f92-46e8-a1e2-be471cf19516"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.565156 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.565200 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.565218 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.565232 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk264\" (UniqueName: \"kubernetes.io/projected/17347053-1f92-46e8-a1e2-be471cf19516-kube-api-access-dk264\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.565282 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17347053-1f92-46e8-a1e2-be471cf19516-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.565315 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.604546 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "17347053-1f92-46e8-a1e2-be471cf19516" (UID: "17347053-1f92-46e8-a1e2-be471cf19516"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.650754 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.662441 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"17347053-1f92-46e8-a1e2-be471cf19516","Type":"ContainerDied","Data":"f9e9233c31d37107a2d6b9e2c76b0d5bfcbfc88ed0bac35e270d482eb23f6d8c"} Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.662935 4876 scope.go:117] "RemoveContainer" containerID="9d9755e38fd10bcc26796d4ab4445627fb53701c36a6fd0b6a7e471def88f7e9" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.662543 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-config-data" (OuterVolumeSpecName: "config-data") pod "17347053-1f92-46e8-a1e2-be471cf19516" (UID: "17347053-1f92-46e8-a1e2-be471cf19516"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.662455 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.665619 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x9tx5" event={"ID":"51b8e947-a71d-4365-a028-7f8d6d184577","Type":"ContainerStarted","Data":"c8cc86a038669528c46ce81fd244531f11ea9f8df5e61a2c7d8c751e9036f9a4"} Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.672214 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6c0e0e38-68d2-45a0-9601-fe2e4f3f9179","Type":"ContainerDied","Data":"aca84be5eed54930d8c202168ed96f5d816f3847da9538b4df3f7e179eff33d9"} Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.672331 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.672363 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.672365 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.672751 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17347053-1f92-46e8-a1e2-be471cf19516-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.681346 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"02e40021-e864-4c53-9522-40fe5719abde","Type":"ContainerStarted","Data":"614c19f51d4d52a286759e882a304f7a5ce9fec5036e52461ce8036cd3ea33d6"} Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.691037 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cb9cf56c6-k7vkg" event={"ID":"9fd26454-5f72-4e52-a3e7-87270228d46d","Type":"ContainerDied","Data":"4b668cfe732b0d00c37d1335159394536244fe6a51656937b48dc42867d7e278"} Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.691085 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b668cfe732b0d00c37d1335159394536244fe6a51656937b48dc42867d7e278" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.715291 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.519653209 podStartE2EDuration="20.715262203s" podCreationTimestamp="2026-03-13 07:59:58 +0000 UTC" firstStartedPulling="2026-03-13 07:59:59.690930663 +0000 UTC m=+1259.361709645" lastFinishedPulling="2026-03-13 08:00:17.886539657 +0000 UTC m=+1277.557318639" observedRunningTime="2026-03-13 08:00:18.701132028 +0000 UTC m=+1278.371911010" watchObservedRunningTime="2026-03-13 08:00:18.715262203 +0000 UTC m=+1278.386041195" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.717633 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66dd85954b-6789x" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.719061 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66dd85954b-6789x" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.737563 4876 scope.go:117] "RemoveContainer" containerID="67d5719f159209f2dad8684db9e7b94b6e9dc28cd3cc9f5d8b5c4ddf4962f041" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.845351 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3c87-account-create-update-mgw9g"] Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.871281 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wxvxg"] Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.895619 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-kl96z"] Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.901765 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.918151 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-75fc-account-create-update-zjt7n"] Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.934981 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-102d-account-create-update-kvlpr"] Mar 13 08:00:18 crc kubenswrapper[4876]: W0313 08:00:18.937632 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fec4b3c_6d95_4ed5_ac57_40da620b201c.slice/crio-3cdff8f312ea649ea97f9ed2433eeb04e9edb77a896a98c9e4e65660c1da11e7 WatchSource:0}: Error finding container 3cdff8f312ea649ea97f9ed2433eeb04e9edb77a896a98c9e4e65660c1da11e7: Status 404 returned error can't find the container with id 3cdff8f312ea649ea97f9ed2433eeb04e9edb77a896a98c9e4e65660c1da11e7 Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.989569 4876 scope.go:117] "RemoveContainer" containerID="f759362668103ca973b7087882f219d00529c8d9e78138c3c9d8b0b0451038b6" Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.993875 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-ovndb-tls-certs\") pod \"9fd26454-5f72-4e52-a3e7-87270228d46d\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.993948 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-httpd-config\") pod \"9fd26454-5f72-4e52-a3e7-87270228d46d\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.993983 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89mmv\" (UniqueName: \"kubernetes.io/projected/9fd26454-5f72-4e52-a3e7-87270228d46d-kube-api-access-89mmv\") pod \"9fd26454-5f72-4e52-a3e7-87270228d46d\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.994050 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-combined-ca-bundle\") pod \"9fd26454-5f72-4e52-a3e7-87270228d46d\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " Mar 13 08:00:18 crc kubenswrapper[4876]: I0313 08:00:18.994287 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-config\") pod \"9fd26454-5f72-4e52-a3e7-87270228d46d\" (UID: \"9fd26454-5f72-4e52-a3e7-87270228d46d\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.011068 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fd26454-5f72-4e52-a3e7-87270228d46d-kube-api-access-89mmv" (OuterVolumeSpecName: "kube-api-access-89mmv") pod "9fd26454-5f72-4e52-a3e7-87270228d46d" (UID: "9fd26454-5f72-4e52-a3e7-87270228d46d"). InnerVolumeSpecName "kube-api-access-89mmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.014969 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.023769 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9fd26454-5f72-4e52-a3e7-87270228d46d" (UID: "9fd26454-5f72-4e52-a3e7-87270228d46d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.035136 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.111094 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.111138 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89mmv\" (UniqueName: \"kubernetes.io/projected/9fd26454-5f72-4e52-a3e7-87270228d46d-kube-api-access-89mmv\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.137844 4876 scope.go:117] "RemoveContainer" containerID="a98aa8db7fa9f9b7ee2f43daf4ea074a3f58593cf14ae19768bbde77177f49bb" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.139841 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" path="/var/lib/kubelet/pods/6c0e0e38-68d2-45a0-9601-fe2e4f3f9179/volumes" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.144896 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 08:00:19 crc kubenswrapper[4876]: E0313 08:00:19.147076 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerName="cinder-api-log" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.147107 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerName="cinder-api-log" Mar 13 08:00:19 crc kubenswrapper[4876]: E0313 08:00:19.147124 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerName="neutron-httpd" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.147132 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerName="neutron-httpd" Mar 13 08:00:19 crc kubenswrapper[4876]: E0313 08:00:19.147153 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerName="cinder-api" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.147161 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerName="cinder-api" Mar 13 08:00:19 crc kubenswrapper[4876]: E0313 08:00:19.147188 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17347053-1f92-46e8-a1e2-be471cf19516" containerName="glance-httpd" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.147196 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="17347053-1f92-46e8-a1e2-be471cf19516" containerName="glance-httpd" Mar 13 08:00:19 crc kubenswrapper[4876]: E0313 08:00:19.147215 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerName="neutron-api" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.147223 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerName="neutron-api" Mar 13 08:00:19 crc kubenswrapper[4876]: E0313 08:00:19.147256 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17347053-1f92-46e8-a1e2-be471cf19516" containerName="glance-log" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.147265 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="17347053-1f92-46e8-a1e2-be471cf19516" containerName="glance-log" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.148785 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerName="neutron-api" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.148832 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="17347053-1f92-46e8-a1e2-be471cf19516" containerName="glance-httpd" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.148852 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerName="cinder-api-log" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.148872 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="17347053-1f92-46e8-a1e2-be471cf19516" containerName="glance-log" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.148888 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c0e0e38-68d2-45a0-9601-fe2e4f3f9179" containerName="cinder-api" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.148898 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" containerName="neutron-httpd" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.151347 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.151492 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.156909 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.157639 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.158630 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.186694 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.199160 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.212056 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.216434 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.232537 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.240445 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.240709 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.319709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.319837 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19442bc3-f0a0-4077-8b11-823aeb39dc3f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.319871 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19442bc3-f0a0-4077-8b11-823aeb39dc3f-logs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.319914 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-config-data\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.319951 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-config-data-custom\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.319982 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320043 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320075 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4def3c5-fa14-4aed-94c0-3374552b56cf-logs\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320107 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320131 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320155 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-scripts\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320183 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4def3c5-fa14-4aed-94c0-3374552b56cf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320214 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320260 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxdvb\" (UniqueName: \"kubernetes.io/projected/19442bc3-f0a0-4077-8b11-823aeb39dc3f-kube-api-access-zxdvb\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320298 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffmwf\" (UniqueName: \"kubernetes.io/projected/e4def3c5-fa14-4aed-94c0-3374552b56cf-kube-api-access-ffmwf\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320337 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.320363 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.399517 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.431942 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432000 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4def3c5-fa14-4aed-94c0-3374552b56cf-logs\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432029 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432046 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432064 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-scripts\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432084 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4def3c5-fa14-4aed-94c0-3374552b56cf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432102 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432125 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxdvb\" (UniqueName: \"kubernetes.io/projected/19442bc3-f0a0-4077-8b11-823aeb39dc3f-kube-api-access-zxdvb\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432153 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffmwf\" (UniqueName: \"kubernetes.io/projected/e4def3c5-fa14-4aed-94c0-3374552b56cf-kube-api-access-ffmwf\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432175 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432192 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432216 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432285 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19442bc3-f0a0-4077-8b11-823aeb39dc3f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432310 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19442bc3-f0a0-4077-8b11-823aeb39dc3f-logs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432335 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-config-data\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432363 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-config-data-custom\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.432399 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.433790 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4def3c5-fa14-4aed-94c0-3374552b56cf-logs\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.435117 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4def3c5-fa14-4aed-94c0-3374552b56cf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.435579 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19442bc3-f0a0-4077-8b11-823aeb39dc3f-logs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.436033 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.438363 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19442bc3-f0a0-4077-8b11-823aeb39dc3f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.449450 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.449896 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.450397 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.451711 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.452427 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.452741 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4def3c5-fa14-4aed-94c0-3374552b56cf-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.454783 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-config-data\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.455306 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.465098 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-scripts\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.470962 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19442bc3-f0a0-4077-8b11-823aeb39dc3f-config-data-custom\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.471356 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffmwf\" (UniqueName: \"kubernetes.io/projected/e4def3c5-fa14-4aed-94c0-3374552b56cf-kube-api-access-ffmwf\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.485411 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxdvb\" (UniqueName: \"kubernetes.io/projected/19442bc3-f0a0-4077-8b11-823aeb39dc3f-kube-api-access-zxdvb\") pod \"cinder-api-0\" (UID: \"19442bc3-f0a0-4077-8b11-823aeb39dc3f\") " pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.534342 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x22nj\" (UniqueName: \"kubernetes.io/projected/b3664dea-7307-496d-aff3-588911d155bb-kube-api-access-x22nj\") pod \"b3664dea-7307-496d-aff3-588911d155bb\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.534461 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"b3664dea-7307-496d-aff3-588911d155bb\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.534580 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-logs\") pod \"b3664dea-7307-496d-aff3-588911d155bb\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.534602 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-config-data\") pod \"b3664dea-7307-496d-aff3-588911d155bb\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.534668 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-httpd-run\") pod \"b3664dea-7307-496d-aff3-588911d155bb\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.534718 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-internal-tls-certs\") pod \"b3664dea-7307-496d-aff3-588911d155bb\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.534762 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-scripts\") pod \"b3664dea-7307-496d-aff3-588911d155bb\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.534780 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-combined-ca-bundle\") pod \"b3664dea-7307-496d-aff3-588911d155bb\" (UID: \"b3664dea-7307-496d-aff3-588911d155bb\") " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.536350 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b3664dea-7307-496d-aff3-588911d155bb" (UID: "b3664dea-7307-496d-aff3-588911d155bb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.536691 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-logs" (OuterVolumeSpecName: "logs") pod "b3664dea-7307-496d-aff3-588911d155bb" (UID: "b3664dea-7307-496d-aff3-588911d155bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.567857 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "b3664dea-7307-496d-aff3-588911d155bb" (UID: "b3664dea-7307-496d-aff3-588911d155bb"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.568114 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3664dea-7307-496d-aff3-588911d155bb-kube-api-access-x22nj" (OuterVolumeSpecName: "kube-api-access-x22nj") pod "b3664dea-7307-496d-aff3-588911d155bb" (UID: "b3664dea-7307-496d-aff3-588911d155bb"). InnerVolumeSpecName "kube-api-access-x22nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.577085 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-scripts" (OuterVolumeSpecName: "scripts") pod "b3664dea-7307-496d-aff3-588911d155bb" (UID: "b3664dea-7307-496d-aff3-588911d155bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.637408 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.637448 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3664dea-7307-496d-aff3-588911d155bb-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.637457 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.637465 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x22nj\" (UniqueName: \"kubernetes.io/projected/b3664dea-7307-496d-aff3-588911d155bb-kube-api-access-x22nj\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.637499 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.651604 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fd26454-5f72-4e52-a3e7-87270228d46d" (UID: "9fd26454-5f72-4e52-a3e7-87270228d46d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.670039 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"e4def3c5-fa14-4aed-94c0-3374552b56cf\") " pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.710965 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wxvxg" event={"ID":"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a","Type":"ContainerStarted","Data":"e51e0aa22794886c2e2d524b9cd03436859216b119ce3353385335fdce520781"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.712799 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556480-8292c" event={"ID":"306b6d6a-7ec0-41f2-b7e6-8974bb873700","Type":"ContainerStarted","Data":"4555ee39a98e78d790cba89dd0f0321a88a7e01669427190e0b33f208846da3e"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.724293 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kl96z" event={"ID":"6e0939f9-b421-4707-861c-b43fae99f079","Type":"ContainerStarted","Data":"2df94f183a914d0bc32415e8ed3f81d41f8785588ac0197e6d732f332e7b5497"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.739560 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.741956 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x9tx5" event={"ID":"51b8e947-a71d-4365-a028-7f8d6d184577","Type":"ContainerStarted","Data":"ea3460326a2d4e125ca36fabec1f6ab8865870c613830208ba3ed170b8678ced"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.747057 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556480-8292c" podStartSLOduration=3.455351475 podStartE2EDuration="19.747026981s" podCreationTimestamp="2026-03-13 08:00:00 +0000 UTC" firstStartedPulling="2026-03-13 08:00:01.649470467 +0000 UTC m=+1261.320249449" lastFinishedPulling="2026-03-13 08:00:17.941145973 +0000 UTC m=+1277.611924955" observedRunningTime="2026-03-13 08:00:19.738040951 +0000 UTC m=+1279.408819933" watchObservedRunningTime="2026-03-13 08:00:19.747026981 +0000 UTC m=+1279.417805963" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.753911 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerStarted","Data":"c470d21baa2b3e7224c1b2bdee256413e3599b9b1b94c809bd7d9a21e8a21dfc"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.756147 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-102d-account-create-update-kvlpr" event={"ID":"3fec4b3c-6d95-4ed5-ac57-40da620b201c","Type":"ContainerStarted","Data":"3cdff8f312ea649ea97f9ed2433eeb04e9edb77a896a98c9e4e65660c1da11e7"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.758809 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.763067 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-config" (OuterVolumeSpecName: "config") pod "9fd26454-5f72-4e52-a3e7-87270228d46d" (UID: "9fd26454-5f72-4e52-a3e7-87270228d46d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.766575 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" event={"ID":"8229f43e-63f9-41a0-8197-4af96c27ee48","Type":"ContainerStarted","Data":"a0c2c9cde78e2ddbd12545a5e60d19d0702f4e03f8e339f2e316ec52a739d6c3"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.766695 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" event={"ID":"8229f43e-63f9-41a0-8197-4af96c27ee48","Type":"ContainerStarted","Data":"1dcaafdd5e14aa29b5db0b3f9c8a41e643cba9e357da0277b26344415aa5b853"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.774551 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3664dea-7307-496d-aff3-588911d155bb","Type":"ContainerDied","Data":"67467ecebda852dfc6fd0eaf843f612964966aa301a9bbc020cad4c65d98e79c"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.774628 4876 scope.go:117] "RemoveContainer" containerID="ac6988d9ae894d1f9bb1fe44270e4e0f5f2765ea43b1cc4650101b6c16699110" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.774785 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.781430 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cb9cf56c6-k7vkg" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.781619 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" event={"ID":"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12","Type":"ContainerStarted","Data":"591f957df911db3a4f899ccf131a26f5a408b1bfb6627884163bb8f3c0a5291d"} Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.783260 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.790672 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3664dea-7307-496d-aff3-588911d155bb" (UID: "b3664dea-7307-496d-aff3-588911d155bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.795207 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9fd26454-5f72-4e52-a3e7-87270228d46d" (UID: "9fd26454-5f72-4e52-a3e7-87270228d46d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.806974 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-x9tx5" podStartSLOduration=8.806812162 podStartE2EDuration="8.806812162s" podCreationTimestamp="2026-03-13 08:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:19.764351415 +0000 UTC m=+1279.435130407" watchObservedRunningTime="2026-03-13 08:00:19.806812162 +0000 UTC m=+1279.477591144" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.826201 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" podStartSLOduration=7.826171043 podStartE2EDuration="7.826171043s" podCreationTimestamp="2026-03-13 08:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:19.793593422 +0000 UTC m=+1279.464372404" watchObservedRunningTime="2026-03-13 08:00:19.826171043 +0000 UTC m=+1279.496950025" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.843363 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-config\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.843404 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.843416 4876 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fd26454-5f72-4e52-a3e7-87270228d46d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.843426 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.843404 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b3664dea-7307-496d-aff3-588911d155bb" (UID: "b3664dea-7307-496d-aff3-588911d155bb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.867002 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-config-data" (OuterVolumeSpecName: "config-data") pod "b3664dea-7307-496d-aff3-588911d155bb" (UID: "b3664dea-7307-496d-aff3-588911d155bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.882623 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.945210 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:19 crc kubenswrapper[4876]: I0313 08:00:19.945704 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3664dea-7307-496d-aff3-588911d155bb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.199203 4876 scope.go:117] "RemoveContainer" containerID="8862cfa4ed45d9337b0f1fd8dc684f0790bbcc8d758836f140c02da1ae2631ea" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.232202 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5cb9cf56c6-k7vkg"] Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.238116 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5cb9cf56c6-k7vkg"] Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.269074 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.289086 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.301215 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 08:00:20 crc kubenswrapper[4876]: E0313 08:00:20.301752 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3664dea-7307-496d-aff3-588911d155bb" containerName="glance-httpd" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.301776 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3664dea-7307-496d-aff3-588911d155bb" containerName="glance-httpd" Mar 13 08:00:20 crc kubenswrapper[4876]: E0313 08:00:20.301820 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3664dea-7307-496d-aff3-588911d155bb" containerName="glance-log" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.301832 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3664dea-7307-496d-aff3-588911d155bb" containerName="glance-log" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.302097 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3664dea-7307-496d-aff3-588911d155bb" containerName="glance-httpd" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.302133 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3664dea-7307-496d-aff3-588911d155bb" containerName="glance-log" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.303274 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.309027 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.311967 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.333916 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.392534 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.473405 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.473904 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd9df755-7613-4abe-a9fc-661e6667eb63-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.473963 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sftkn\" (UniqueName: \"kubernetes.io/projected/dd9df755-7613-4abe-a9fc-661e6667eb63-kube-api-access-sftkn\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.474010 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.474037 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd9df755-7613-4abe-a9fc-661e6667eb63-logs\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.474099 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.474133 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.474181 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.576470 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.576530 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.576652 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.576772 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.576850 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd9df755-7613-4abe-a9fc-661e6667eb63-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.576898 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sftkn\" (UniqueName: \"kubernetes.io/projected/dd9df755-7613-4abe-a9fc-661e6667eb63-kube-api-access-sftkn\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.576941 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.576969 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd9df755-7613-4abe-a9fc-661e6667eb63-logs\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.577928 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.583828 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd9df755-7613-4abe-a9fc-661e6667eb63-logs\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.584746 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.587678 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dd9df755-7613-4abe-a9fc-661e6667eb63-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.595818 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.602330 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sftkn\" (UniqueName: \"kubernetes.io/projected/dd9df755-7613-4abe-a9fc-661e6667eb63-kube-api-access-sftkn\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.615702 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.620921 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd9df755-7613-4abe-a9fc-661e6667eb63-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: W0313 08:00:20.629998 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4def3c5_fa14_4aed_94c0_3374552b56cf.slice/crio-d4edd74c77535fd249fab4e5a29b0d5ebc1b5c8d1ee82cf71f54aa0eda37e62f WatchSource:0}: Error finding container d4edd74c77535fd249fab4e5a29b0d5ebc1b5c8d1ee82cf71f54aa0eda37e62f: Status 404 returned error can't find the container with id d4edd74c77535fd249fab4e5a29b0d5ebc1b5c8d1ee82cf71f54aa0eda37e62f Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.651398 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.688091 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"dd9df755-7613-4abe-a9fc-661e6667eb63\") " pod="openstack/glance-default-internal-api-0" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.809953 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4def3c5-fa14-4aed-94c0-3374552b56cf","Type":"ContainerStarted","Data":"d4edd74c77535fd249fab4e5a29b0d5ebc1b5c8d1ee82cf71f54aa0eda37e62f"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.814225 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wxvxg" event={"ID":"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a","Type":"ContainerStarted","Data":"4401cf7ec14b2aba19ace026bf6e00c3b0a74f90c86153ad7eb9df1e8ffb6cb3"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.818489 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerStarted","Data":"78f4e644173b7f3302894f0ad9053e390bf7f744062e95401dc0f8243906be39"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.821279 4876 generic.go:334] "Generic (PLEG): container finished" podID="8229f43e-63f9-41a0-8197-4af96c27ee48" containerID="a0c2c9cde78e2ddbd12545a5e60d19d0702f4e03f8e339f2e316ec52a739d6c3" exitCode=0 Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.821340 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" event={"ID":"8229f43e-63f9-41a0-8197-4af96c27ee48","Type":"ContainerDied","Data":"a0c2c9cde78e2ddbd12545a5e60d19d0702f4e03f8e339f2e316ec52a739d6c3"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.824046 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"19442bc3-f0a0-4077-8b11-823aeb39dc3f","Type":"ContainerStarted","Data":"9118e0f88708fc6c6f8416efa415b1919a86f978d2f780e23f521143325476ed"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.825884 4876 generic.go:334] "Generic (PLEG): container finished" podID="306b6d6a-7ec0-41f2-b7e6-8974bb873700" containerID="4555ee39a98e78d790cba89dd0f0321a88a7e01669427190e0b33f208846da3e" exitCode=0 Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.825965 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556480-8292c" event={"ID":"306b6d6a-7ec0-41f2-b7e6-8974bb873700","Type":"ContainerDied","Data":"4555ee39a98e78d790cba89dd0f0321a88a7e01669427190e0b33f208846da3e"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.833188 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-102d-account-create-update-kvlpr" event={"ID":"3fec4b3c-6d95-4ed5-ac57-40da620b201c","Type":"ContainerStarted","Data":"dd54696bdd240b84feca773d507f3bd64ce6cfddb93affa36fdfda799f0c1f0f"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.837560 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-wxvxg" podStartSLOduration=8.837539202 podStartE2EDuration="8.837539202s" podCreationTimestamp="2026-03-13 08:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:20.829301411 +0000 UTC m=+1280.500080393" watchObservedRunningTime="2026-03-13 08:00:20.837539202 +0000 UTC m=+1280.508318184" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.841804 4876 generic.go:334] "Generic (PLEG): container finished" podID="6e0939f9-b421-4707-861c-b43fae99f079" containerID="766fe4c0e9d9606bd4a06d79c54564e2fd59dcec56c37a379437bb0a4821939a" exitCode=0 Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.841966 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kl96z" event={"ID":"6e0939f9-b421-4707-861c-b43fae99f079","Type":"ContainerDied","Data":"766fe4c0e9d9606bd4a06d79c54564e2fd59dcec56c37a379437bb0a4821939a"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.850689 4876 generic.go:334] "Generic (PLEG): container finished" podID="51b8e947-a71d-4365-a028-7f8d6d184577" containerID="ea3460326a2d4e125ca36fabec1f6ab8865870c613830208ba3ed170b8678ced" exitCode=0 Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.850754 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x9tx5" event={"ID":"51b8e947-a71d-4365-a028-7f8d6d184577","Type":"ContainerDied","Data":"ea3460326a2d4e125ca36fabec1f6ab8865870c613830208ba3ed170b8678ced"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.863739 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" event={"ID":"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12","Type":"ContainerStarted","Data":"b2c87db8f35c5801cef873bcdfa601e4673a38239f4e43e1199f1892dca44fe2"} Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.930641 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-102d-account-create-update-kvlpr" podStartSLOduration=8.930617152 podStartE2EDuration="8.930617152s" podCreationTimestamp="2026-03-13 08:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:20.922411253 +0000 UTC m=+1280.593190245" watchObservedRunningTime="2026-03-13 08:00:20.930617152 +0000 UTC m=+1280.601396134" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.943306 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" podStartSLOduration=8.943283056 podStartE2EDuration="8.943283056s" podCreationTimestamp="2026-03-13 08:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:20.938471612 +0000 UTC m=+1280.609250614" watchObservedRunningTime="2026-03-13 08:00:20.943283056 +0000 UTC m=+1280.614062038" Mar 13 08:00:20 crc kubenswrapper[4876]: I0313 08:00:20.958052 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.051086 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17347053-1f92-46e8-a1e2-be471cf19516" path="/var/lib/kubelet/pods/17347053-1f92-46e8-a1e2-be471cf19516/volumes" Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.053184 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fd26454-5f72-4e52-a3e7-87270228d46d" path="/var/lib/kubelet/pods/9fd26454-5f72-4e52-a3e7-87270228d46d/volumes" Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.054162 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3664dea-7307-496d-aff3-588911d155bb" path="/var/lib/kubelet/pods/b3664dea-7307-496d-aff3-588911d155bb/volumes" Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.646625 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.881757 4876 generic.go:334] "Generic (PLEG): container finished" podID="5c8d7466-ceca-4879-9263-9c9eeb0d3a7a" containerID="4401cf7ec14b2aba19ace026bf6e00c3b0a74f90c86153ad7eb9df1e8ffb6cb3" exitCode=0 Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.881857 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wxvxg" event={"ID":"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a","Type":"ContainerDied","Data":"4401cf7ec14b2aba19ace026bf6e00c3b0a74f90c86153ad7eb9df1e8ffb6cb3"} Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.887229 4876 generic.go:334] "Generic (PLEG): container finished" podID="3fec4b3c-6d95-4ed5-ac57-40da620b201c" containerID="dd54696bdd240b84feca773d507f3bd64ce6cfddb93affa36fdfda799f0c1f0f" exitCode=0 Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.887320 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-102d-account-create-update-kvlpr" event={"ID":"3fec4b3c-6d95-4ed5-ac57-40da620b201c","Type":"ContainerDied","Data":"dd54696bdd240b84feca773d507f3bd64ce6cfddb93affa36fdfda799f0c1f0f"} Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.891987 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd9df755-7613-4abe-a9fc-661e6667eb63","Type":"ContainerStarted","Data":"8789b8f386b0ca0d424850d41ed16aaaeb2dbe5cabeb655a9309dab6f9cb786a"} Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.898768 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerStarted","Data":"dcca7debd5b8cd22cc56204986eb23b1a779774e0c8759ef5f99a62a405aa6b8"} Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.919686 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"19442bc3-f0a0-4077-8b11-823aeb39dc3f","Type":"ContainerStarted","Data":"2e84cf8033c8b46c5fe16c0de244a436146b6160e051b3e1d72f8ad60577ae5b"} Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.923672 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4def3c5-fa14-4aed-94c0-3374552b56cf","Type":"ContainerStarted","Data":"d8d470765eff132da900c546db493ba2089e6f2ec0f0a041d434bcc3d54b984b"} Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.934455 4876 generic.go:334] "Generic (PLEG): container finished" podID="7ecb0fb1-66c7-400e-90cb-7f5bbe847f12" containerID="b2c87db8f35c5801cef873bcdfa601e4673a38239f4e43e1199f1892dca44fe2" exitCode=0 Mar 13 08:00:21 crc kubenswrapper[4876]: I0313 08:00:21.935301 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" event={"ID":"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12","Type":"ContainerDied","Data":"b2c87db8f35c5801cef873bcdfa601e4673a38239f4e43e1199f1892dca44fe2"} Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.411573 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.537716 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8229f43e-63f9-41a0-8197-4af96c27ee48-operator-scripts\") pod \"8229f43e-63f9-41a0-8197-4af96c27ee48\" (UID: \"8229f43e-63f9-41a0-8197-4af96c27ee48\") " Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.538274 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6kgp\" (UniqueName: \"kubernetes.io/projected/8229f43e-63f9-41a0-8197-4af96c27ee48-kube-api-access-w6kgp\") pod \"8229f43e-63f9-41a0-8197-4af96c27ee48\" (UID: \"8229f43e-63f9-41a0-8197-4af96c27ee48\") " Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.538756 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8229f43e-63f9-41a0-8197-4af96c27ee48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8229f43e-63f9-41a0-8197-4af96c27ee48" (UID: "8229f43e-63f9-41a0-8197-4af96c27ee48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.539377 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8229f43e-63f9-41a0-8197-4af96c27ee48-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.544643 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8229f43e-63f9-41a0-8197-4af96c27ee48-kube-api-access-w6kgp" (OuterVolumeSpecName: "kube-api-access-w6kgp") pod "8229f43e-63f9-41a0-8197-4af96c27ee48" (UID: "8229f43e-63f9-41a0-8197-4af96c27ee48"). InnerVolumeSpecName "kube-api-access-w6kgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.643289 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6kgp\" (UniqueName: \"kubernetes.io/projected/8229f43e-63f9-41a0-8197-4af96c27ee48-kube-api-access-w6kgp\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.736219 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.736986 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556480-8292c" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.750084 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.848223 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhjgz\" (UniqueName: \"kubernetes.io/projected/51b8e947-a71d-4365-a028-7f8d6d184577-kube-api-access-zhjgz\") pod \"51b8e947-a71d-4365-a028-7f8d6d184577\" (UID: \"51b8e947-a71d-4365-a028-7f8d6d184577\") " Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.849665 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0939f9-b421-4707-861c-b43fae99f079-operator-scripts\") pod \"6e0939f9-b421-4707-861c-b43fae99f079\" (UID: \"6e0939f9-b421-4707-861c-b43fae99f079\") " Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.849758 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51b8e947-a71d-4365-a028-7f8d6d184577-operator-scripts\") pod \"51b8e947-a71d-4365-a028-7f8d6d184577\" (UID: \"51b8e947-a71d-4365-a028-7f8d6d184577\") " Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.849963 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lptf9\" (UniqueName: \"kubernetes.io/projected/306b6d6a-7ec0-41f2-b7e6-8974bb873700-kube-api-access-lptf9\") pod \"306b6d6a-7ec0-41f2-b7e6-8974bb873700\" (UID: \"306b6d6a-7ec0-41f2-b7e6-8974bb873700\") " Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.850012 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tn74\" (UniqueName: \"kubernetes.io/projected/6e0939f9-b421-4707-861c-b43fae99f079-kube-api-access-6tn74\") pod \"6e0939f9-b421-4707-861c-b43fae99f079\" (UID: \"6e0939f9-b421-4707-861c-b43fae99f079\") " Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.851250 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e0939f9-b421-4707-861c-b43fae99f079-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6e0939f9-b421-4707-861c-b43fae99f079" (UID: "6e0939f9-b421-4707-861c-b43fae99f079"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.852008 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51b8e947-a71d-4365-a028-7f8d6d184577-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51b8e947-a71d-4365-a028-7f8d6d184577" (UID: "51b8e947-a71d-4365-a028-7f8d6d184577"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.863071 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/306b6d6a-7ec0-41f2-b7e6-8974bb873700-kube-api-access-lptf9" (OuterVolumeSpecName: "kube-api-access-lptf9") pod "306b6d6a-7ec0-41f2-b7e6-8974bb873700" (UID: "306b6d6a-7ec0-41f2-b7e6-8974bb873700"). InnerVolumeSpecName "kube-api-access-lptf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.865541 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b8e947-a71d-4365-a028-7f8d6d184577-kube-api-access-zhjgz" (OuterVolumeSpecName: "kube-api-access-zhjgz") pod "51b8e947-a71d-4365-a028-7f8d6d184577" (UID: "51b8e947-a71d-4365-a028-7f8d6d184577"). InnerVolumeSpecName "kube-api-access-zhjgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.868731 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0939f9-b421-4707-861c-b43fae99f079-kube-api-access-6tn74" (OuterVolumeSpecName: "kube-api-access-6tn74") pod "6e0939f9-b421-4707-861c-b43fae99f079" (UID: "6e0939f9-b421-4707-861c-b43fae99f079"). InnerVolumeSpecName "kube-api-access-6tn74". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.955824 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e0939f9-b421-4707-861c-b43fae99f079-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.955875 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51b8e947-a71d-4365-a028-7f8d6d184577-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.955889 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lptf9\" (UniqueName: \"kubernetes.io/projected/306b6d6a-7ec0-41f2-b7e6-8974bb873700-kube-api-access-lptf9\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.955907 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tn74\" (UniqueName: \"kubernetes.io/projected/6e0939f9-b421-4707-861c-b43fae99f079-kube-api-access-6tn74\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.955918 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhjgz\" (UniqueName: \"kubernetes.io/projected/51b8e947-a71d-4365-a028-7f8d6d184577-kube-api-access-zhjgz\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.984937 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x9tx5" event={"ID":"51b8e947-a71d-4365-a028-7f8d6d184577","Type":"ContainerDied","Data":"c8cc86a038669528c46ce81fd244531f11ea9f8df5e61a2c7d8c751e9036f9a4"} Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.985015 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8cc86a038669528c46ce81fd244531f11ea9f8df5e61a2c7d8c751e9036f9a4" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.985107 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x9tx5" Mar 13 08:00:22 crc kubenswrapper[4876]: I0313 08:00:22.995577 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd9df755-7613-4abe-a9fc-661e6667eb63","Type":"ContainerStarted","Data":"a9ae87903568b5f745d16ddefa9f56aeab4abbbd283eba788a9c8864a9d0f3ba"} Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:22.998327 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" event={"ID":"8229f43e-63f9-41a0-8197-4af96c27ee48","Type":"ContainerDied","Data":"1dcaafdd5e14aa29b5db0b3f9c8a41e643cba9e357da0277b26344415aa5b853"} Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:22.998370 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dcaafdd5e14aa29b5db0b3f9c8a41e643cba9e357da0277b26344415aa5b853" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:22.998511 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3c87-account-create-update-mgw9g" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.034216 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"19442bc3-f0a0-4077-8b11-823aeb39dc3f","Type":"ContainerStarted","Data":"21e8c70ae69e84aea2e4bcbfcdc4343f7b5996af2e5d0c71e895a01b923a0b2d"} Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.035915 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.103375 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.10334877 podStartE2EDuration="4.10334877s" podCreationTimestamp="2026-03-13 08:00:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:23.077060236 +0000 UTC m=+1282.747839218" watchObservedRunningTime="2026-03-13 08:00:23.10334877 +0000 UTC m=+1282.774127752" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.109786 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4def3c5-fa14-4aed-94c0-3374552b56cf","Type":"ContainerStarted","Data":"c4e9a6e5e54342028b58776c3446c8f6c124e8ceba3a8df5259fd0b2dc57ed61"} Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.111001 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556480-8292c" event={"ID":"306b6d6a-7ec0-41f2-b7e6-8974bb873700","Type":"ContainerDied","Data":"ce54e944e19f28d93ba865a5d9e7939cbcd27d4e386d9c7760f5ec68eb51242c"} Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.111052 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce54e944e19f28d93ba865a5d9e7939cbcd27d4e386d9c7760f5ec68eb51242c" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.111156 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556480-8292c" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.125738 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kl96z" event={"ID":"6e0939f9-b421-4707-861c-b43fae99f079","Type":"ContainerDied","Data":"2df94f183a914d0bc32415e8ed3f81d41f8785588ac0197e6d732f332e7b5497"} Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.125805 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2df94f183a914d0bc32415e8ed3f81d41f8785588ac0197e6d732f332e7b5497" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.125940 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kl96z" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.696707 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.736986 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.736950214 podStartE2EDuration="4.736950214s" podCreationTimestamp="2026-03-13 08:00:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:23.139773638 +0000 UTC m=+1282.810552620" watchObservedRunningTime="2026-03-13 08:00:23.736950214 +0000 UTC m=+1283.407729206" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.790805 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fec4b3c-6d95-4ed5-ac57-40da620b201c-operator-scripts\") pod \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\" (UID: \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\") " Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.790980 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvsrr\" (UniqueName: \"kubernetes.io/projected/3fec4b3c-6d95-4ed5-ac57-40da620b201c-kube-api-access-rvsrr\") pod \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\" (UID: \"3fec4b3c-6d95-4ed5-ac57-40da620b201c\") " Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.792197 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fec4b3c-6d95-4ed5-ac57-40da620b201c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3fec4b3c-6d95-4ed5-ac57-40da620b201c" (UID: "3fec4b3c-6d95-4ed5-ac57-40da620b201c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.794117 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fec4b3c-6d95-4ed5-ac57-40da620b201c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.808501 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fec4b3c-6d95-4ed5-ac57-40da620b201c-kube-api-access-rvsrr" (OuterVolumeSpecName: "kube-api-access-rvsrr") pod "3fec4b3c-6d95-4ed5-ac57-40da620b201c" (UID: "3fec4b3c-6d95-4ed5-ac57-40da620b201c"). InnerVolumeSpecName "kube-api-access-rvsrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.818206 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.830372 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.895278 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-operator-scripts\") pod \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\" (UID: \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\") " Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.895357 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmqgd\" (UniqueName: \"kubernetes.io/projected/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-kube-api-access-nmqgd\") pod \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\" (UID: \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\") " Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.895459 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jv9w\" (UniqueName: \"kubernetes.io/projected/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-kube-api-access-2jv9w\") pod \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\" (UID: \"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12\") " Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.895500 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-operator-scripts\") pod \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\" (UID: \"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a\") " Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.895948 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvsrr\" (UniqueName: \"kubernetes.io/projected/3fec4b3c-6d95-4ed5-ac57-40da620b201c-kube-api-access-rvsrr\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.896434 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c8d7466-ceca-4879-9263-9c9eeb0d3a7a" (UID: "5c8d7466-ceca-4879-9263-9c9eeb0d3a7a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.896827 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ecb0fb1-66c7-400e-90cb-7f5bbe847f12" (UID: "7ecb0fb1-66c7-400e-90cb-7f5bbe847f12"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.907586 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-kube-api-access-2jv9w" (OuterVolumeSpecName: "kube-api-access-2jv9w") pod "7ecb0fb1-66c7-400e-90cb-7f5bbe847f12" (UID: "7ecb0fb1-66c7-400e-90cb-7f5bbe847f12"). InnerVolumeSpecName "kube-api-access-2jv9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.908917 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-kube-api-access-nmqgd" (OuterVolumeSpecName: "kube-api-access-nmqgd") pod "5c8d7466-ceca-4879-9263-9c9eeb0d3a7a" (UID: "5c8d7466-ceca-4879-9263-9c9eeb0d3a7a"). InnerVolumeSpecName "kube-api-access-nmqgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:23 crc kubenswrapper[4876]: I0313 08:00:23.988787 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556474-24kvs"] Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.000340 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.000384 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmqgd\" (UniqueName: \"kubernetes.io/projected/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-kube-api-access-nmqgd\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.000397 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jv9w\" (UniqueName: \"kubernetes.io/projected/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12-kube-api-access-2jv9w\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.000408 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.005310 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556474-24kvs"] Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.159527 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" event={"ID":"7ecb0fb1-66c7-400e-90cb-7f5bbe847f12","Type":"ContainerDied","Data":"591f957df911db3a4f899ccf131a26f5a408b1bfb6627884163bb8f3c0a5291d"} Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.159581 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="591f957df911db3a4f899ccf131a26f5a408b1bfb6627884163bb8f3c0a5291d" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.159676 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-75fc-account-create-update-zjt7n" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.167675 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wxvxg" event={"ID":"5c8d7466-ceca-4879-9263-9c9eeb0d3a7a","Type":"ContainerDied","Data":"e51e0aa22794886c2e2d524b9cd03436859216b119ce3353385335fdce520781"} Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.167732 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e51e0aa22794886c2e2d524b9cd03436859216b119ce3353385335fdce520781" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.167840 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wxvxg" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.173706 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-102d-account-create-update-kvlpr" event={"ID":"3fec4b3c-6d95-4ed5-ac57-40da620b201c","Type":"ContainerDied","Data":"3cdff8f312ea649ea97f9ed2433eeb04e9edb77a896a98c9e4e65660c1da11e7"} Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.173769 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cdff8f312ea649ea97f9ed2433eeb04e9edb77a896a98c9e4e65660c1da11e7" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.173863 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-102d-account-create-update-kvlpr" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.193912 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dd9df755-7613-4abe-a9fc-661e6667eb63","Type":"ContainerStarted","Data":"77e93e8ab383ac215a31815349ea4796ba16914133e707b3e94e2b8db25ca2cf"} Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.210836 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="ceilometer-central-agent" containerID="cri-o://c470d21baa2b3e7224c1b2bdee256413e3599b9b1b94c809bd7d9a21e8a21dfc" gracePeriod=30 Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.210944 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerStarted","Data":"cb1c39bcd6bcd4f303c36078f194ac7f5958f4b970935b4a99c5dc3d2fc3c662"} Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.211517 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.211568 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="proxy-httpd" containerID="cri-o://cb1c39bcd6bcd4f303c36078f194ac7f5958f4b970935b4a99c5dc3d2fc3c662" gracePeriod=30 Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.211612 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="sg-core" containerID="cri-o://dcca7debd5b8cd22cc56204986eb23b1a779774e0c8759ef5f99a62a405aa6b8" gracePeriod=30 Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.211653 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="ceilometer-notification-agent" containerID="cri-o://78f4e644173b7f3302894f0ad9053e390bf7f744062e95401dc0f8243906be39" gracePeriod=30 Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.260547 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.260511983 podStartE2EDuration="4.260511983s" podCreationTimestamp="2026-03-13 08:00:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:24.248869767 +0000 UTC m=+1283.919648749" watchObservedRunningTime="2026-03-13 08:00:24.260511983 +0000 UTC m=+1283.931290985" Mar 13 08:00:24 crc kubenswrapper[4876]: I0313 08:00:24.306883 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.347491737 podStartE2EDuration="16.306852108s" podCreationTimestamp="2026-03-13 08:00:08 +0000 UTC" firstStartedPulling="2026-03-13 08:00:17.509986866 +0000 UTC m=+1277.180765858" lastFinishedPulling="2026-03-13 08:00:23.469347237 +0000 UTC m=+1283.140126229" observedRunningTime="2026-03-13 08:00:24.29261576 +0000 UTC m=+1283.963394742" watchObservedRunningTime="2026-03-13 08:00:24.306852108 +0000 UTC m=+1283.977631090" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.053727 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2a6f4f-87ef-4109-a53d-c48a82c094aa" path="/var/lib/kubelet/pods/5e2a6f4f-87ef-4109-a53d-c48a82c094aa/volumes" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.149534 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66dd85954b-6789x" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.222192 4876 generic.go:334] "Generic (PLEG): container finished" podID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerID="b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8" exitCode=137 Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.222279 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66dd85954b-6789x" event={"ID":"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163","Type":"ContainerDied","Data":"b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8"} Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.222331 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66dd85954b-6789x" event={"ID":"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163","Type":"ContainerDied","Data":"fc61cb1cc6bb2cd2db65514e555a97acf5d04e079d61d9a8f67f8b321e75e6fb"} Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.222354 4876 scope.go:117] "RemoveContainer" containerID="645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.222367 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66dd85954b-6789x" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.231931 4876 generic.go:334] "Generic (PLEG): container finished" podID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerID="dcca7debd5b8cd22cc56204986eb23b1a779774e0c8759ef5f99a62a405aa6b8" exitCode=2 Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.231976 4876 generic.go:334] "Generic (PLEG): container finished" podID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerID="78f4e644173b7f3302894f0ad9053e390bf7f744062e95401dc0f8243906be39" exitCode=0 Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.231988 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerDied","Data":"dcca7debd5b8cd22cc56204986eb23b1a779774e0c8759ef5f99a62a405aa6b8"} Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.232028 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerDied","Data":"78f4e644173b7f3302894f0ad9053e390bf7f744062e95401dc0f8243906be39"} Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.330836 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66zhr\" (UniqueName: \"kubernetes.io/projected/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-kube-api-access-66zhr\") pod \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.331051 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-combined-ca-bundle\") pod \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.331087 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-scripts\") pod \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.331127 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-secret-key\") pod \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.331175 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-tls-certs\") pod \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.331225 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-logs\") pod \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.331299 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-config-data\") pod \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\" (UID: \"6ed427db-e9a0-4b8f-b142-a2ab5bd0c163\") " Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.335585 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-logs" (OuterVolumeSpecName: "logs") pod "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" (UID: "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.339443 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" (UID: "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.343370 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-kube-api-access-66zhr" (OuterVolumeSpecName: "kube-api-access-66zhr") pod "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" (UID: "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163"). InnerVolumeSpecName "kube-api-access-66zhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.406394 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" (UID: "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.410103 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-config-data" (OuterVolumeSpecName: "config-data") pod "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" (UID: "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.412520 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" (UID: "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.415185 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-scripts" (OuterVolumeSpecName: "scripts") pod "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" (UID: "6ed427db-e9a0-4b8f-b142-a2ab5bd0c163"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.428846 4876 scope.go:117] "RemoveContainer" containerID="b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.434277 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66zhr\" (UniqueName: \"kubernetes.io/projected/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-kube-api-access-66zhr\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.434311 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.434324 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.434333 4876 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.434343 4876 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.434351 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.434363 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.458436 4876 scope.go:117] "RemoveContainer" containerID="645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74" Mar 13 08:00:25 crc kubenswrapper[4876]: E0313 08:00:25.459043 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74\": container with ID starting with 645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74 not found: ID does not exist" containerID="645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.459096 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74"} err="failed to get container status \"645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74\": rpc error: code = NotFound desc = could not find container \"645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74\": container with ID starting with 645c2fb14bbbd9d8044380a22b6ddc46cf7ad41872a1393502387ef0a17a4a74 not found: ID does not exist" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.459127 4876 scope.go:117] "RemoveContainer" containerID="b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8" Mar 13 08:00:25 crc kubenswrapper[4876]: E0313 08:00:25.462165 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8\": container with ID starting with b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8 not found: ID does not exist" containerID="b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.462194 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8"} err="failed to get container status \"b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8\": rpc error: code = NotFound desc = could not find container \"b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8\": container with ID starting with b6d10cee92b3365435966c64fad4de3247c1258f9f13dbd074cf94e9535238c8 not found: ID does not exist" Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.561028 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66dd85954b-6789x"] Mar 13 08:00:25 crc kubenswrapper[4876]: I0313 08:00:25.575384 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66dd85954b-6789x"] Mar 13 08:00:26 crc kubenswrapper[4876]: I0313 08:00:26.246108 4876 generic.go:334] "Generic (PLEG): container finished" podID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerID="c470d21baa2b3e7224c1b2bdee256413e3599b9b1b94c809bd7d9a21e8a21dfc" exitCode=0 Mar 13 08:00:26 crc kubenswrapper[4876]: I0313 08:00:26.246174 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerDied","Data":"c470d21baa2b3e7224c1b2bdee256413e3599b9b1b94c809bd7d9a21e8a21dfc"} Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.049360 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" path="/var/lib/kubelet/pods/6ed427db-e9a0-4b8f-b142-a2ab5bd0c163/volumes" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.763923 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzp92"] Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.764826 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon-log" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.764844 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon-log" Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.764866 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="306b6d6a-7ec0-41f2-b7e6-8974bb873700" containerName="oc" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.764874 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="306b6d6a-7ec0-41f2-b7e6-8974bb873700" containerName="oc" Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.764883 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ecb0fb1-66c7-400e-90cb-7f5bbe847f12" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.764891 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ecb0fb1-66c7-400e-90cb-7f5bbe847f12" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.764910 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8229f43e-63f9-41a0-8197-4af96c27ee48" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.764918 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8229f43e-63f9-41a0-8197-4af96c27ee48" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.764933 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0939f9-b421-4707-861c-b43fae99f079" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.764940 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0939f9-b421-4707-861c-b43fae99f079" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.764954 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8d7466-ceca-4879-9263-9c9eeb0d3a7a" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.764961 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8d7466-ceca-4879-9263-9c9eeb0d3a7a" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.764975 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b8e947-a71d-4365-a028-7f8d6d184577" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.764982 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b8e947-a71d-4365-a028-7f8d6d184577" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.765007 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fec4b3c-6d95-4ed5-ac57-40da620b201c" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765014 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fec4b3c-6d95-4ed5-ac57-40da620b201c" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: E0313 08:00:27.765038 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765046 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765270 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765281 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="306b6d6a-7ec0-41f2-b7e6-8974bb873700" containerName="oc" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765291 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8229f43e-63f9-41a0-8197-4af96c27ee48" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765305 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fec4b3c-6d95-4ed5-ac57-40da620b201c" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765314 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8d7466-ceca-4879-9263-9c9eeb0d3a7a" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765321 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ecb0fb1-66c7-400e-90cb-7f5bbe847f12" containerName="mariadb-account-create-update" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765338 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed427db-e9a0-4b8f-b142-a2ab5bd0c163" containerName="horizon-log" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765358 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b8e947-a71d-4365-a028-7f8d6d184577" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.765368 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0939f9-b421-4707-861c-b43fae99f079" containerName="mariadb-database-create" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.766254 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.775876 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-7thc2" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.776262 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.777813 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.796470 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzp92"] Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.884540 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-config-data\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.884629 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b676n\" (UniqueName: \"kubernetes.io/projected/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-kube-api-access-b676n\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.884671 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.884779 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-scripts\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.986253 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-scripts\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.986343 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-config-data\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.986399 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b676n\" (UniqueName: \"kubernetes.io/projected/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-kube-api-access-b676n\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.986435 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.993987 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:27 crc kubenswrapper[4876]: I0313 08:00:27.994532 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-config-data\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:28 crc kubenswrapper[4876]: I0313 08:00:28.007722 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-scripts\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:28 crc kubenswrapper[4876]: I0313 08:00:28.016923 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b676n\" (UniqueName: \"kubernetes.io/projected/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-kube-api-access-b676n\") pod \"nova-cell0-conductor-db-sync-jzp92\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:28 crc kubenswrapper[4876]: I0313 08:00:28.102738 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:28 crc kubenswrapper[4876]: I0313 08:00:28.623568 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzp92"] Mar 13 08:00:29 crc kubenswrapper[4876]: I0313 08:00:29.297455 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jzp92" event={"ID":"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce","Type":"ContainerStarted","Data":"15be2b43cd37369841d686460f4ea691a1f68df8d014c3a2ea8eaec33d2a0d2a"} Mar 13 08:00:29 crc kubenswrapper[4876]: I0313 08:00:29.884553 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 08:00:29 crc kubenswrapper[4876]: I0313 08:00:29.884669 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 08:00:29 crc kubenswrapper[4876]: I0313 08:00:29.939156 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 08:00:29 crc kubenswrapper[4876]: I0313 08:00:29.960105 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 08:00:30 crc kubenswrapper[4876]: I0313 08:00:30.320341 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 08:00:30 crc kubenswrapper[4876]: I0313 08:00:30.320533 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 08:00:30 crc kubenswrapper[4876]: I0313 08:00:30.958971 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:30 crc kubenswrapper[4876]: I0313 08:00:30.959062 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:31 crc kubenswrapper[4876]: I0313 08:00:31.012310 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:31 crc kubenswrapper[4876]: I0313 08:00:31.016463 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:31 crc kubenswrapper[4876]: I0313 08:00:31.330596 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:31 crc kubenswrapper[4876]: I0313 08:00:31.330640 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:32 crc kubenswrapper[4876]: I0313 08:00:32.528993 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 08:00:32 crc kubenswrapper[4876]: I0313 08:00:32.690395 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 08:00:32 crc kubenswrapper[4876]: I0313 08:00:32.691003 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 08:00:32 crc kubenswrapper[4876]: I0313 08:00:32.922016 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 08:00:33 crc kubenswrapper[4876]: I0313 08:00:33.349873 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 08:00:33 crc kubenswrapper[4876]: I0313 08:00:33.349947 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 08:00:34 crc kubenswrapper[4876]: I0313 08:00:34.138138 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:34 crc kubenswrapper[4876]: I0313 08:00:34.142155 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 08:00:38 crc kubenswrapper[4876]: I0313 08:00:38.824874 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 08:00:41 crc kubenswrapper[4876]: I0313 08:00:41.452025 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jzp92" event={"ID":"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce","Type":"ContainerStarted","Data":"2a7835d82325345577bbddaa180c61e60521ec132ac9d86add5588dfc18edce0"} Mar 13 08:00:51 crc kubenswrapper[4876]: I0313 08:00:51.561866 4876 generic.go:334] "Generic (PLEG): container finished" podID="d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" containerID="2a7835d82325345577bbddaa180c61e60521ec132ac9d86add5588dfc18edce0" exitCode=0 Mar 13 08:00:51 crc kubenswrapper[4876]: I0313 08:00:51.561988 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jzp92" event={"ID":"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce","Type":"ContainerDied","Data":"2a7835d82325345577bbddaa180c61e60521ec132ac9d86add5588dfc18edce0"} Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.000681 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.164129 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-config-data\") pod \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.164275 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-combined-ca-bundle\") pod \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.164402 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-scripts\") pod \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.164460 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b676n\" (UniqueName: \"kubernetes.io/projected/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-kube-api-access-b676n\") pod \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\" (UID: \"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce\") " Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.173708 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-kube-api-access-b676n" (OuterVolumeSpecName: "kube-api-access-b676n") pod "d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" (UID: "d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce"). InnerVolumeSpecName "kube-api-access-b676n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.185666 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-scripts" (OuterVolumeSpecName: "scripts") pod "d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" (UID: "d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.207986 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-config-data" (OuterVolumeSpecName: "config-data") pod "d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" (UID: "d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.224392 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" (UID: "d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.267418 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.267464 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.267483 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b676n\" (UniqueName: \"kubernetes.io/projected/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-kube-api-access-b676n\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.267502 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.587768 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jzp92" event={"ID":"d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce","Type":"ContainerDied","Data":"15be2b43cd37369841d686460f4ea691a1f68df8d014c3a2ea8eaec33d2a0d2a"} Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.588193 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15be2b43cd37369841d686460f4ea691a1f68df8d014c3a2ea8eaec33d2a0d2a" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.587897 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jzp92" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.699146 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 08:00:53 crc kubenswrapper[4876]: E0313 08:00:53.699829 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" containerName="nova-cell0-conductor-db-sync" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.699873 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" containerName="nova-cell0-conductor-db-sync" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.700162 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" containerName="nova-cell0-conductor-db-sync" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.701398 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.707826 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-7thc2" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.708045 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.758812 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.779185 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cld7c\" (UniqueName: \"kubernetes.io/projected/a77871ec-0f52-4d09-a3e2-664d14c9adbf-kube-api-access-cld7c\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.779419 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a77871ec-0f52-4d09-a3e2-664d14c9adbf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.779465 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a77871ec-0f52-4d09-a3e2-664d14c9adbf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.881914 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a77871ec-0f52-4d09-a3e2-664d14c9adbf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.881991 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a77871ec-0f52-4d09-a3e2-664d14c9adbf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.882089 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cld7c\" (UniqueName: \"kubernetes.io/projected/a77871ec-0f52-4d09-a3e2-664d14c9adbf-kube-api-access-cld7c\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.887067 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a77871ec-0f52-4d09-a3e2-664d14c9adbf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.891338 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a77871ec-0f52-4d09-a3e2-664d14c9adbf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:53 crc kubenswrapper[4876]: I0313 08:00:53.907808 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cld7c\" (UniqueName: \"kubernetes.io/projected/a77871ec-0f52-4d09-a3e2-664d14c9adbf-kube-api-access-cld7c\") pod \"nova-cell0-conductor-0\" (UID: \"a77871ec-0f52-4d09-a3e2-664d14c9adbf\") " pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.074464 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.545917 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 08:00:54 crc kubenswrapper[4876]: W0313 08:00:54.552624 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda77871ec_0f52_4d09_a3e2_664d14c9adbf.slice/crio-5fb21c07dd357805e4621ebbf732d5cb3353eb37eda740c61384bef813612ba6 WatchSource:0}: Error finding container 5fb21c07dd357805e4621ebbf732d5cb3353eb37eda740c61384bef813612ba6: Status 404 returned error can't find the container with id 5fb21c07dd357805e4621ebbf732d5cb3353eb37eda740c61384bef813612ba6 Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.603377 4876 generic.go:334] "Generic (PLEG): container finished" podID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerID="cb1c39bcd6bcd4f303c36078f194ac7f5958f4b970935b4a99c5dc3d2fc3c662" exitCode=137 Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.603563 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerDied","Data":"cb1c39bcd6bcd4f303c36078f194ac7f5958f4b970935b4a99c5dc3d2fc3c662"} Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.603658 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"018ec9da-cec8-4c3c-a5b7-c652ef88de04","Type":"ContainerDied","Data":"45afb5de0caf327d6aea6bb3637e47878a0d5667121b8a79ff5ee3a3f6da7c85"} Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.603674 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45afb5de0caf327d6aea6bb3637e47878a0d5667121b8a79ff5ee3a3f6da7c85" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.605556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a77871ec-0f52-4d09-a3e2-664d14c9adbf","Type":"ContainerStarted","Data":"5fb21c07dd357805e4621ebbf732d5cb3353eb37eda740c61384bef813612ba6"} Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.608389 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.698520 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-config-data\") pod \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.698884 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-run-httpd\") pod \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.699085 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-combined-ca-bundle\") pod \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.699292 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-scripts\") pod \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.699421 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz6hk\" (UniqueName: \"kubernetes.io/projected/018ec9da-cec8-4c3c-a5b7-c652ef88de04-kube-api-access-jz6hk\") pod \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.699611 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-log-httpd\") pod \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.699746 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-sg-core-conf-yaml\") pod \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\" (UID: \"018ec9da-cec8-4c3c-a5b7-c652ef88de04\") " Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.699971 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "018ec9da-cec8-4c3c-a5b7-c652ef88de04" (UID: "018ec9da-cec8-4c3c-a5b7-c652ef88de04"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.700079 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "018ec9da-cec8-4c3c-a5b7-c652ef88de04" (UID: "018ec9da-cec8-4c3c-a5b7-c652ef88de04"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.700632 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.700907 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/018ec9da-cec8-4c3c-a5b7-c652ef88de04-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.705425 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-scripts" (OuterVolumeSpecName: "scripts") pod "018ec9da-cec8-4c3c-a5b7-c652ef88de04" (UID: "018ec9da-cec8-4c3c-a5b7-c652ef88de04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.706155 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/018ec9da-cec8-4c3c-a5b7-c652ef88de04-kube-api-access-jz6hk" (OuterVolumeSpecName: "kube-api-access-jz6hk") pod "018ec9da-cec8-4c3c-a5b7-c652ef88de04" (UID: "018ec9da-cec8-4c3c-a5b7-c652ef88de04"). InnerVolumeSpecName "kube-api-access-jz6hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.742533 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "018ec9da-cec8-4c3c-a5b7-c652ef88de04" (UID: "018ec9da-cec8-4c3c-a5b7-c652ef88de04"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.788431 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "018ec9da-cec8-4c3c-a5b7-c652ef88de04" (UID: "018ec9da-cec8-4c3c-a5b7-c652ef88de04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.802927 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.802967 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.802980 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz6hk\" (UniqueName: \"kubernetes.io/projected/018ec9da-cec8-4c3c-a5b7-c652ef88de04-kube-api-access-jz6hk\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.802991 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.820812 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-config-data" (OuterVolumeSpecName: "config-data") pod "018ec9da-cec8-4c3c-a5b7-c652ef88de04" (UID: "018ec9da-cec8-4c3c-a5b7-c652ef88de04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:00:54 crc kubenswrapper[4876]: I0313 08:00:54.904537 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ec9da-cec8-4c3c-a5b7-c652ef88de04-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.622697 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.624520 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a77871ec-0f52-4d09-a3e2-664d14c9adbf","Type":"ContainerStarted","Data":"d1cc98b8a35c2bd2e5339224cab32c9840caa2170c1c79f78bc32d107c0c3cdf"} Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.624561 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.654775 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.65474134 podStartE2EDuration="2.65474134s" podCreationTimestamp="2026-03-13 08:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:00:55.64218248 +0000 UTC m=+1315.312961482" watchObservedRunningTime="2026-03-13 08:00:55.65474134 +0000 UTC m=+1315.325520332" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.678085 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.685829 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.697791 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:55 crc kubenswrapper[4876]: E0313 08:00:55.698464 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="ceilometer-notification-agent" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.698490 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="ceilometer-notification-agent" Mar 13 08:00:55 crc kubenswrapper[4876]: E0313 08:00:55.698528 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="proxy-httpd" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.698538 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="proxy-httpd" Mar 13 08:00:55 crc kubenswrapper[4876]: E0313 08:00:55.698561 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="sg-core" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.698570 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="sg-core" Mar 13 08:00:55 crc kubenswrapper[4876]: E0313 08:00:55.698578 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="ceilometer-central-agent" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.698585 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="ceilometer-central-agent" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.698758 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="proxy-httpd" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.698788 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="ceilometer-central-agent" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.698798 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="ceilometer-notification-agent" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.698806 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" containerName="sg-core" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.700890 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.704198 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.704678 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.717359 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.827690 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.827751 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.827823 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frvhr\" (UniqueName: \"kubernetes.io/projected/3ddfd44e-033d-49c2-b174-c239e223388b-kube-api-access-frvhr\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.827922 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-log-httpd\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.828276 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-config-data\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.828348 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-run-httpd\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.828650 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-scripts\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.930885 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-log-httpd\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.930988 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-config-data\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.931010 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-run-httpd\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.931050 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-scripts\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.931082 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.931100 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.931146 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frvhr\" (UniqueName: \"kubernetes.io/projected/3ddfd44e-033d-49c2-b174-c239e223388b-kube-api-access-frvhr\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.932086 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-log-httpd\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.933433 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-run-httpd\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.939091 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-scripts\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.939567 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.941067 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-config-data\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.941086 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:55 crc kubenswrapper[4876]: I0313 08:00:55.955144 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frvhr\" (UniqueName: \"kubernetes.io/projected/3ddfd44e-033d-49c2-b174-c239e223388b-kube-api-access-frvhr\") pod \"ceilometer-0\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " pod="openstack/ceilometer-0" Mar 13 08:00:56 crc kubenswrapper[4876]: I0313 08:00:56.023409 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:00:56 crc kubenswrapper[4876]: W0313 08:00:56.385970 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ddfd44e_033d_49c2_b174_c239e223388b.slice/crio-a66fa78401219287c1244f5322785ee0935873e52a4f522bbc3781f8eb34afe2 WatchSource:0}: Error finding container a66fa78401219287c1244f5322785ee0935873e52a4f522bbc3781f8eb34afe2: Status 404 returned error can't find the container with id a66fa78401219287c1244f5322785ee0935873e52a4f522bbc3781f8eb34afe2 Mar 13 08:00:56 crc kubenswrapper[4876]: I0313 08:00:56.391392 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:00:56 crc kubenswrapper[4876]: I0313 08:00:56.637087 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerStarted","Data":"a66fa78401219287c1244f5322785ee0935873e52a4f522bbc3781f8eb34afe2"} Mar 13 08:00:57 crc kubenswrapper[4876]: I0313 08:00:57.052433 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="018ec9da-cec8-4c3c-a5b7-c652ef88de04" path="/var/lib/kubelet/pods/018ec9da-cec8-4c3c-a5b7-c652ef88de04/volumes" Mar 13 08:00:57 crc kubenswrapper[4876]: I0313 08:00:57.660589 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerStarted","Data":"3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491"} Mar 13 08:00:58 crc kubenswrapper[4876]: I0313 08:00:58.674393 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerStarted","Data":"9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207"} Mar 13 08:00:58 crc kubenswrapper[4876]: I0313 08:00:58.675130 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerStarted","Data":"2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330"} Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.114406 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.659332 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lrrb8"] Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.661685 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.665808 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.666086 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.673737 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lrrb8"] Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.728229 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-scripts\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.728317 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.728381 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-config-data\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.728403 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzlvc\" (UniqueName: \"kubernetes.io/projected/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-kube-api-access-mzlvc\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.830118 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-scripts\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.830214 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.830316 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-config-data\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.830342 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzlvc\" (UniqueName: \"kubernetes.io/projected/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-kube-api-access-mzlvc\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.844438 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-config-data\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.845456 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.845559 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-scripts\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.866888 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzlvc\" (UniqueName: \"kubernetes.io/projected/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-kube-api-access-mzlvc\") pod \"nova-cell0-cell-mapping-lrrb8\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.867531 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.869511 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.875018 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.876795 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.932032 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.932553 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.933020 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwfwg\" (UniqueName: \"kubernetes.io/projected/b304c04a-a07f-487d-8b8e-7379fa339016-kube-api-access-hwfwg\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:00:59 crc kubenswrapper[4876]: I0313 08:00:59.981918 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.002480 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.003899 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.011936 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.034927 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwfwg\" (UniqueName: \"kubernetes.io/projected/b304c04a-a07f-487d-8b8e-7379fa339016-kube-api-access-hwfwg\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.035337 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.035443 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.038341 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.052459 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.056042 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.079899 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwfwg\" (UniqueName: \"kubernetes.io/projected/b304c04a-a07f-487d-8b8e-7379fa339016-kube-api-access-hwfwg\") pod \"nova-cell1-novncproxy-0\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.082150 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.084987 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.102748 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.141376 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b297r\" (UniqueName: \"kubernetes.io/projected/92a20643-48f7-4b68-a270-358ee507a8cf-kube-api-access-b297r\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.141540 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-config-data\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.141582 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.185475 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.255333 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.255547 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.255584 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-config-data\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.255609 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.255746 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd5a59b-fa90-4d42-96bf-bc318917490f-logs\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.255792 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b297r\" (UniqueName: \"kubernetes.io/projected/92a20643-48f7-4b68-a270-358ee507a8cf-kube-api-access-b297r\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.255825 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvvq2\" (UniqueName: \"kubernetes.io/projected/cdd5a59b-fa90-4d42-96bf-bc318917490f-kube-api-access-cvvq2\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.255873 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-config-data\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.257117 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.261830 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.264392 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.269700 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.273030 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.300742 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29556481-xrkdk"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.301973 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-config-data\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.302384 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.310180 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b297r\" (UniqueName: \"kubernetes.io/projected/92a20643-48f7-4b68-a270-358ee507a8cf-kube-api-access-b297r\") pod \"nova-scheduler-0\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.335926 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.336485 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556481-xrkdk"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.359602 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4adea-ea85-40c6-ad78-d19e36ded44e-logs\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.359675 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvvq2\" (UniqueName: \"kubernetes.io/projected/cdd5a59b-fa90-4d42-96bf-bc318917490f-kube-api-access-cvvq2\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.359711 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-config-data\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.359760 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.359784 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-config-data\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.359835 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.359912 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kms8\" (UniqueName: \"kubernetes.io/projected/58c4adea-ea85-40c6-ad78-d19e36ded44e-kube-api-access-6kms8\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.359938 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd5a59b-fa90-4d42-96bf-bc318917490f-logs\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.360407 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd5a59b-fa90-4d42-96bf-bc318917490f-logs\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.369642 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-47hc5"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.371780 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.385102 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-47hc5"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.414111 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvvq2\" (UniqueName: \"kubernetes.io/projected/cdd5a59b-fa90-4d42-96bf-bc318917490f-kube-api-access-cvvq2\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.414593 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-config-data\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.415110 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.480968 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-config-data\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.481522 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.482615 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.482777 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-fernet-keys\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.482871 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-svc\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.482907 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kms8\" (UniqueName: \"kubernetes.io/projected/58c4adea-ea85-40c6-ad78-d19e36ded44e-kube-api-access-6kms8\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.483049 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4adea-ea85-40c6-ad78-d19e36ded44e-logs\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.483084 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-combined-ca-bundle\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.483180 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.483263 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.483314 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-config-data\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.483393 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f77x\" (UniqueName: \"kubernetes.io/projected/44b18fd5-828d-41ce-a174-b7b116d9e8bb-kube-api-access-2f77x\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.483429 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7p89\" (UniqueName: \"kubernetes.io/projected/a91beee0-58e6-4d98-8000-e4e6bbee83a8-kube-api-access-t7p89\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.483469 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-config\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.485714 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.486554 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4adea-ea85-40c6-ad78-d19e36ded44e-logs\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.492305 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-config-data\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.495738 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.507074 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kms8\" (UniqueName: \"kubernetes.io/projected/58c4adea-ea85-40c6-ad78-d19e36ded44e-kube-api-access-6kms8\") pod \"nova-metadata-0\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.589800 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-svc\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.589918 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-combined-ca-bundle\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.589954 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.589976 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.590008 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f77x\" (UniqueName: \"kubernetes.io/projected/44b18fd5-828d-41ce-a174-b7b116d9e8bb-kube-api-access-2f77x\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.590030 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7p89\" (UniqueName: \"kubernetes.io/projected/a91beee0-58e6-4d98-8000-e4e6bbee83a8-kube-api-access-t7p89\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.590052 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-config\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.590078 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-config-data\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.590108 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.590168 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-fernet-keys\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.593711 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-config\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.594648 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.594977 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.595498 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-svc\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.595657 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.596471 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-fernet-keys\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.599198 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-config-data\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.608732 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-combined-ca-bundle\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.615401 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7p89\" (UniqueName: \"kubernetes.io/projected/a91beee0-58e6-4d98-8000-e4e6bbee83a8-kube-api-access-t7p89\") pod \"keystone-cron-29556481-xrkdk\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.616218 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f77x\" (UniqueName: \"kubernetes.io/projected/44b18fd5-828d-41ce-a174-b7b116d9e8bb-kube-api-access-2f77x\") pod \"dnsmasq-dns-757b4f8459-47hc5\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.654940 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.736929 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.750196 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.940993 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.951767 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lrrb8"] Mar 13 08:01:00 crc kubenswrapper[4876]: I0313 08:01:00.999482 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.226905 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5st8q"] Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.229904 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.234006 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.234491 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.252578 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5st8q"] Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.308993 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.309873 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.309998 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-config-data\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.310051 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8fsb\" (UniqueName: \"kubernetes.io/projected/812ab290-71fc-4db3-94af-6daffce9da81-kube-api-access-d8fsb\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.310140 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-scripts\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.412444 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.412855 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-config-data\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.412883 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8fsb\" (UniqueName: \"kubernetes.io/projected/812ab290-71fc-4db3-94af-6daffce9da81-kube-api-access-d8fsb\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.412956 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-scripts\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.421955 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.422345 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-config-data\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.424791 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-scripts\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.437422 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8fsb\" (UniqueName: \"kubernetes.io/projected/812ab290-71fc-4db3-94af-6daffce9da81-kube-api-access-d8fsb\") pod \"nova-cell1-conductor-db-sync-5st8q\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.546081 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.554530 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-47hc5"] Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.585172 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.733560 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556481-xrkdk"] Mar 13 08:01:01 crc kubenswrapper[4876]: W0313 08:01:01.770445 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda91beee0_58e6_4d98_8000_e4e6bbee83a8.slice/crio-09076561c36e95d81cb58e7705ce2acca533b4114793ee3a538f99f81b28632f WatchSource:0}: Error finding container 09076561c36e95d81cb58e7705ce2acca533b4114793ee3a538f99f81b28632f: Status 404 returned error can't find the container with id 09076561c36e95d81cb58e7705ce2acca533b4114793ee3a538f99f81b28632f Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.799865 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerStarted","Data":"ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8"} Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.801653 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.807002 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92a20643-48f7-4b68-a270-358ee507a8cf","Type":"ContainerStarted","Data":"dce0b4f8bf70050cda0d0905fb110097e62c2919f3d25296d8fb8fb407230f53"} Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.817665 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" event={"ID":"44b18fd5-828d-41ce-a174-b7b116d9e8bb","Type":"ContainerStarted","Data":"fad686ba62c961f45d857292ad8e4e3be9b37e75f988126bed23af4d3816eeb8"} Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.833014 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lrrb8" event={"ID":"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4","Type":"ContainerStarted","Data":"58ded0c1d939e2e1aaaa3bcfa7a860776ba705aba34f9551cc684c9b8cdaabd6"} Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.833078 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lrrb8" event={"ID":"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4","Type":"ContainerStarted","Data":"7a9b1d38cbd3bce6e59f86fcda313ebce52cde04fe26f8b6444742a3d2393f17"} Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.858843 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b304c04a-a07f-487d-8b8e-7379fa339016","Type":"ContainerStarted","Data":"cdc52b62b1066b080f9d4f963e23b223a15fee5ab341dea30ca2930433c2bf82"} Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.875713 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.608315876 podStartE2EDuration="6.875616018s" podCreationTimestamp="2026-03-13 08:00:55 +0000 UTC" firstStartedPulling="2026-03-13 08:00:56.392189396 +0000 UTC m=+1316.062968378" lastFinishedPulling="2026-03-13 08:01:00.659489538 +0000 UTC m=+1320.330268520" observedRunningTime="2026-03-13 08:01:01.832802351 +0000 UTC m=+1321.503581333" watchObservedRunningTime="2026-03-13 08:01:01.875616018 +0000 UTC m=+1321.546395000" Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.878775 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4adea-ea85-40c6-ad78-d19e36ded44e","Type":"ContainerStarted","Data":"e74e842eb7b73be742bd7aaa14b458797dc37d881d7ad0217d474930677eec53"} Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.883587 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdd5a59b-fa90-4d42-96bf-bc318917490f","Type":"ContainerStarted","Data":"cbd4317f4dd6b931c50ea7a70dda3895078781f344069a51b985e45b6ae51861"} Mar 13 08:01:01 crc kubenswrapper[4876]: I0313 08:01:01.899354 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lrrb8" podStartSLOduration=2.899322461 podStartE2EDuration="2.899322461s" podCreationTimestamp="2026-03-13 08:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:01.876008589 +0000 UTC m=+1321.546787591" watchObservedRunningTime="2026-03-13 08:01:01.899322461 +0000 UTC m=+1321.570101443" Mar 13 08:01:02 crc kubenswrapper[4876]: I0313 08:01:02.171349 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5st8q"] Mar 13 08:01:02 crc kubenswrapper[4876]: I0313 08:01:02.932290 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556481-xrkdk" event={"ID":"a91beee0-58e6-4d98-8000-e4e6bbee83a8","Type":"ContainerStarted","Data":"8173dcb79c37ef84b225c405741d651deaa5ad3e3956c77251aa82bbbad6acac"} Mar 13 08:01:02 crc kubenswrapper[4876]: I0313 08:01:02.932775 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556481-xrkdk" event={"ID":"a91beee0-58e6-4d98-8000-e4e6bbee83a8","Type":"ContainerStarted","Data":"09076561c36e95d81cb58e7705ce2acca533b4114793ee3a538f99f81b28632f"} Mar 13 08:01:02 crc kubenswrapper[4876]: I0313 08:01:02.938292 4876 generic.go:334] "Generic (PLEG): container finished" podID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" containerID="38752a93dd3eeba29aae494d45d90ad83eed588344764fd59b8a0700883cf4c6" exitCode=0 Mar 13 08:01:02 crc kubenswrapper[4876]: I0313 08:01:02.938361 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" event={"ID":"44b18fd5-828d-41ce-a174-b7b116d9e8bb","Type":"ContainerDied","Data":"38752a93dd3eeba29aae494d45d90ad83eed588344764fd59b8a0700883cf4c6"} Mar 13 08:01:02 crc kubenswrapper[4876]: I0313 08:01:02.959180 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5st8q" event={"ID":"812ab290-71fc-4db3-94af-6daffce9da81","Type":"ContainerStarted","Data":"d3e8eecbdc105c24b74b8b8b513b6288e98ca1cd156c41fa36e38a8cc9ef88d4"} Mar 13 08:01:02 crc kubenswrapper[4876]: I0313 08:01:02.959279 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5st8q" event={"ID":"812ab290-71fc-4db3-94af-6daffce9da81","Type":"ContainerStarted","Data":"18a152bbeec98549cdbd6dcce0ebbdae51fa0de518d792585b479fcce0345939"} Mar 13 08:01:02 crc kubenswrapper[4876]: I0313 08:01:02.963977 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29556481-xrkdk" podStartSLOduration=2.963949557 podStartE2EDuration="2.963949557s" podCreationTimestamp="2026-03-13 08:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:02.95296036 +0000 UTC m=+1322.623739342" watchObservedRunningTime="2026-03-13 08:01:02.963949557 +0000 UTC m=+1322.634728539" Mar 13 08:01:03 crc kubenswrapper[4876]: I0313 08:01:03.030780 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-5st8q" podStartSLOduration=2.030745914 podStartE2EDuration="2.030745914s" podCreationTimestamp="2026-03-13 08:01:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:02.99627356 +0000 UTC m=+1322.667052552" watchObservedRunningTime="2026-03-13 08:01:03.030745914 +0000 UTC m=+1322.701524896" Mar 13 08:01:03 crc kubenswrapper[4876]: I0313 08:01:03.952017 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:03 crc kubenswrapper[4876]: I0313 08:01:03.967701 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:01:03 crc kubenswrapper[4876]: I0313 08:01:03.988397 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" event={"ID":"44b18fd5-828d-41ce-a174-b7b116d9e8bb","Type":"ContainerStarted","Data":"54d08cc86dafab14b9d50bc909b298cea46af95983be2fded4bf1c1c28665452"} Mar 13 08:01:03 crc kubenswrapper[4876]: I0313 08:01:03.988866 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:04 crc kubenswrapper[4876]: I0313 08:01:04.031223 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" podStartSLOduration=4.031184266 podStartE2EDuration="4.031184266s" podCreationTimestamp="2026-03-13 08:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:04.016306411 +0000 UTC m=+1323.687085413" watchObservedRunningTime="2026-03-13 08:01:04.031184266 +0000 UTC m=+1323.701963268" Mar 13 08:01:06 crc kubenswrapper[4876]: I0313 08:01:06.032930 4876 generic.go:334] "Generic (PLEG): container finished" podID="a91beee0-58e6-4d98-8000-e4e6bbee83a8" containerID="8173dcb79c37ef84b225c405741d651deaa5ad3e3956c77251aa82bbbad6acac" exitCode=0 Mar 13 08:01:06 crc kubenswrapper[4876]: I0313 08:01:06.033036 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556481-xrkdk" event={"ID":"a91beee0-58e6-4d98-8000-e4e6bbee83a8","Type":"ContainerDied","Data":"8173dcb79c37ef84b225c405741d651deaa5ad3e3956c77251aa82bbbad6acac"} Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.405272 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.492358 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7p89\" (UniqueName: \"kubernetes.io/projected/a91beee0-58e6-4d98-8000-e4e6bbee83a8-kube-api-access-t7p89\") pod \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.492598 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-fernet-keys\") pod \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.492770 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-config-data\") pod \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.493569 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-combined-ca-bundle\") pod \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\" (UID: \"a91beee0-58e6-4d98-8000-e4e6bbee83a8\") " Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.497627 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a91beee0-58e6-4d98-8000-e4e6bbee83a8-kube-api-access-t7p89" (OuterVolumeSpecName: "kube-api-access-t7p89") pod "a91beee0-58e6-4d98-8000-e4e6bbee83a8" (UID: "a91beee0-58e6-4d98-8000-e4e6bbee83a8"). InnerVolumeSpecName "kube-api-access-t7p89". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.498594 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a91beee0-58e6-4d98-8000-e4e6bbee83a8" (UID: "a91beee0-58e6-4d98-8000-e4e6bbee83a8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.525498 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a91beee0-58e6-4d98-8000-e4e6bbee83a8" (UID: "a91beee0-58e6-4d98-8000-e4e6bbee83a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.558179 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-config-data" (OuterVolumeSpecName: "config-data") pod "a91beee0-58e6-4d98-8000-e4e6bbee83a8" (UID: "a91beee0-58e6-4d98-8000-e4e6bbee83a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.596887 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7p89\" (UniqueName: \"kubernetes.io/projected/a91beee0-58e6-4d98-8000-e4e6bbee83a8-kube-api-access-t7p89\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.596933 4876 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.596947 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:07 crc kubenswrapper[4876]: I0313 08:01:07.596958 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91beee0-58e6-4d98-8000-e4e6bbee83a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.065209 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b304c04a-a07f-487d-8b8e-7379fa339016","Type":"ContainerStarted","Data":"337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4"} Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.065404 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="b304c04a-a07f-487d-8b8e-7379fa339016" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4" gracePeriod=30 Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.081063 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4adea-ea85-40c6-ad78-d19e36ded44e","Type":"ContainerStarted","Data":"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75"} Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.081648 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4adea-ea85-40c6-ad78-d19e36ded44e","Type":"ContainerStarted","Data":"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c"} Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.081377 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerName="nova-metadata-metadata" containerID="cri-o://8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75" gracePeriod=30 Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.081142 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerName="nova-metadata-log" containerID="cri-o://2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c" gracePeriod=30 Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.095366 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdd5a59b-fa90-4d42-96bf-bc318917490f","Type":"ContainerStarted","Data":"c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a"} Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.095420 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdd5a59b-fa90-4d42-96bf-bc318917490f","Type":"ContainerStarted","Data":"e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6"} Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.104079 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556481-xrkdk" event={"ID":"a91beee0-58e6-4d98-8000-e4e6bbee83a8","Type":"ContainerDied","Data":"09076561c36e95d81cb58e7705ce2acca533b4114793ee3a538f99f81b28632f"} Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.104129 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556481-xrkdk" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.104148 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09076561c36e95d81cb58e7705ce2acca533b4114793ee3a538f99f81b28632f" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.114878 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.395385041 podStartE2EDuration="9.114839778s" podCreationTimestamp="2026-03-13 08:00:59 +0000 UTC" firstStartedPulling="2026-03-13 08:01:00.9788052 +0000 UTC m=+1320.649584182" lastFinishedPulling="2026-03-13 08:01:06.698259937 +0000 UTC m=+1326.369038919" observedRunningTime="2026-03-13 08:01:08.09378528 +0000 UTC m=+1327.764564292" watchObservedRunningTime="2026-03-13 08:01:08.114839778 +0000 UTC m=+1327.785618760" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.127731 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92a20643-48f7-4b68-a270-358ee507a8cf","Type":"ContainerStarted","Data":"3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0"} Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.140677 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.002548655 podStartE2EDuration="8.140650139s" podCreationTimestamp="2026-03-13 08:01:00 +0000 UTC" firstStartedPulling="2026-03-13 08:01:01.56647873 +0000 UTC m=+1321.237257712" lastFinishedPulling="2026-03-13 08:01:06.704580214 +0000 UTC m=+1326.375359196" observedRunningTime="2026-03-13 08:01:08.118900392 +0000 UTC m=+1327.789679374" watchObservedRunningTime="2026-03-13 08:01:08.140650139 +0000 UTC m=+1327.811429121" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.157306 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.798565816 podStartE2EDuration="9.157278204s" podCreationTimestamp="2026-03-13 08:00:59 +0000 UTC" firstStartedPulling="2026-03-13 08:01:01.339634902 +0000 UTC m=+1321.010413884" lastFinishedPulling="2026-03-13 08:01:06.69834729 +0000 UTC m=+1326.369126272" observedRunningTime="2026-03-13 08:01:08.144080505 +0000 UTC m=+1327.814859497" watchObservedRunningTime="2026-03-13 08:01:08.157278204 +0000 UTC m=+1327.828057186" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.177957 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.615739428 podStartE2EDuration="9.177927091s" podCreationTimestamp="2026-03-13 08:00:59 +0000 UTC" firstStartedPulling="2026-03-13 08:01:01.135967162 +0000 UTC m=+1320.806746144" lastFinishedPulling="2026-03-13 08:01:06.698154825 +0000 UTC m=+1326.368933807" observedRunningTime="2026-03-13 08:01:08.170510604 +0000 UTC m=+1327.841289586" watchObservedRunningTime="2026-03-13 08:01:08.177927091 +0000 UTC m=+1327.848706063" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.743730 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.819376 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4adea-ea85-40c6-ad78-d19e36ded44e-logs\") pod \"58c4adea-ea85-40c6-ad78-d19e36ded44e\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.819453 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-combined-ca-bundle\") pod \"58c4adea-ea85-40c6-ad78-d19e36ded44e\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.819526 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-config-data\") pod \"58c4adea-ea85-40c6-ad78-d19e36ded44e\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.819612 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kms8\" (UniqueName: \"kubernetes.io/projected/58c4adea-ea85-40c6-ad78-d19e36ded44e-kube-api-access-6kms8\") pod \"58c4adea-ea85-40c6-ad78-d19e36ded44e\" (UID: \"58c4adea-ea85-40c6-ad78-d19e36ded44e\") " Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.820014 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c4adea-ea85-40c6-ad78-d19e36ded44e-logs" (OuterVolumeSpecName: "logs") pod "58c4adea-ea85-40c6-ad78-d19e36ded44e" (UID: "58c4adea-ea85-40c6-ad78-d19e36ded44e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.820796 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4adea-ea85-40c6-ad78-d19e36ded44e-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.830180 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c4adea-ea85-40c6-ad78-d19e36ded44e-kube-api-access-6kms8" (OuterVolumeSpecName: "kube-api-access-6kms8") pod "58c4adea-ea85-40c6-ad78-d19e36ded44e" (UID: "58c4adea-ea85-40c6-ad78-d19e36ded44e"). InnerVolumeSpecName "kube-api-access-6kms8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.900529 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-config-data" (OuterVolumeSpecName: "config-data") pod "58c4adea-ea85-40c6-ad78-d19e36ded44e" (UID: "58c4adea-ea85-40c6-ad78-d19e36ded44e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.915429 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58c4adea-ea85-40c6-ad78-d19e36ded44e" (UID: "58c4adea-ea85-40c6-ad78-d19e36ded44e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.923564 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.923612 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4adea-ea85-40c6-ad78-d19e36ded44e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:08 crc kubenswrapper[4876]: I0313 08:01:08.923661 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kms8\" (UniqueName: \"kubernetes.io/projected/58c4adea-ea85-40c6-ad78-d19e36ded44e-kube-api-access-6kms8\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.143764 4876 generic.go:334] "Generic (PLEG): container finished" podID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerID="8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75" exitCode=0 Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.143805 4876 generic.go:334] "Generic (PLEG): container finished" podID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerID="2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c" exitCode=143 Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.144152 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4adea-ea85-40c6-ad78-d19e36ded44e","Type":"ContainerDied","Data":"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75"} Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.144363 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4adea-ea85-40c6-ad78-d19e36ded44e","Type":"ContainerDied","Data":"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c"} Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.144384 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4adea-ea85-40c6-ad78-d19e36ded44e","Type":"ContainerDied","Data":"e74e842eb7b73be742bd7aaa14b458797dc37d881d7ad0217d474930677eec53"} Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.144409 4876 scope.go:117] "RemoveContainer" containerID="8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.144694 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.184891 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.191902 4876 scope.go:117] "RemoveContainer" containerID="2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.197760 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.213058 4876 scope.go:117] "RemoveContainer" containerID="8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75" Mar 13 08:01:09 crc kubenswrapper[4876]: E0313 08:01:09.213760 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75\": container with ID starting with 8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75 not found: ID does not exist" containerID="8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.213810 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75"} err="failed to get container status \"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75\": rpc error: code = NotFound desc = could not find container \"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75\": container with ID starting with 8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75 not found: ID does not exist" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.213842 4876 scope.go:117] "RemoveContainer" containerID="2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c" Mar 13 08:01:09 crc kubenswrapper[4876]: E0313 08:01:09.214294 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c\": container with ID starting with 2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c not found: ID does not exist" containerID="2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.214321 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c"} err="failed to get container status \"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c\": rpc error: code = NotFound desc = could not find container \"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c\": container with ID starting with 2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c not found: ID does not exist" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.214353 4876 scope.go:117] "RemoveContainer" containerID="8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.214714 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75"} err="failed to get container status \"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75\": rpc error: code = NotFound desc = could not find container \"8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75\": container with ID starting with 8e1a733116dc40767036dee945d853c961fb86e548d1a120268ec5a0b18e7f75 not found: ID does not exist" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.214775 4876 scope.go:117] "RemoveContainer" containerID="2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.215065 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c"} err="failed to get container status \"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c\": rpc error: code = NotFound desc = could not find container \"2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c\": container with ID starting with 2f394adc7c547bc130311d0f28490794cdef6b8b795aff56c2d8c62c44c6875c not found: ID does not exist" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.220939 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:09 crc kubenswrapper[4876]: E0313 08:01:09.221533 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerName="nova-metadata-metadata" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.221554 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerName="nova-metadata-metadata" Mar 13 08:01:09 crc kubenswrapper[4876]: E0313 08:01:09.221593 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerName="nova-metadata-log" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.221601 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerName="nova-metadata-log" Mar 13 08:01:09 crc kubenswrapper[4876]: E0313 08:01:09.221612 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a91beee0-58e6-4d98-8000-e4e6bbee83a8" containerName="keystone-cron" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.221618 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a91beee0-58e6-4d98-8000-e4e6bbee83a8" containerName="keystone-cron" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.221847 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a91beee0-58e6-4d98-8000-e4e6bbee83a8" containerName="keystone-cron" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.221909 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerName="nova-metadata-log" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.221926 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" containerName="nova-metadata-metadata" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.223180 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.226702 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.227308 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.241862 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.332723 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.332791 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.332896 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x948s\" (UniqueName: \"kubernetes.io/projected/9d96cdbd-9db3-4099-b06e-3199437682ef-kube-api-access-x948s\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.333014 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d96cdbd-9db3-4099-b06e-3199437682ef-logs\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.333105 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-config-data\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.435058 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d96cdbd-9db3-4099-b06e-3199437682ef-logs\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.435210 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-config-data\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.435290 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.435325 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.435388 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x948s\" (UniqueName: \"kubernetes.io/projected/9d96cdbd-9db3-4099-b06e-3199437682ef-kube-api-access-x948s\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.435596 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d96cdbd-9db3-4099-b06e-3199437682ef-logs\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.442676 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.450900 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.451271 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-config-data\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.452673 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x948s\" (UniqueName: \"kubernetes.io/projected/9d96cdbd-9db3-4099-b06e-3199437682ef-kube-api-access-x948s\") pod \"nova-metadata-0\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " pod="openstack/nova-metadata-0" Mar 13 08:01:09 crc kubenswrapper[4876]: I0313 08:01:09.548559 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.037969 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.157295 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d96cdbd-9db3-4099-b06e-3199437682ef","Type":"ContainerStarted","Data":"f09cf9953fe8b802aa40612e9ab01eb50f59f0fa26f41364f1db0105c7fcb368"} Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.264778 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.338916 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.339548 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.371835 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.486304 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.486656 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.752843 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.844982 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-7rqrl"] Mar 13 08:01:10 crc kubenswrapper[4876]: I0313 08:01:10.845376 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" podUID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerName="dnsmasq-dns" containerID="cri-o://148e33290be9b0f22e06f90a2510aab995fd0af4f7d3acd9c95fa7efa28dd22b" gracePeriod=10 Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.054041 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c4adea-ea85-40c6-ad78-d19e36ded44e" path="/var/lib/kubelet/pods/58c4adea-ea85-40c6-ad78-d19e36ded44e/volumes" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.184202 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d96cdbd-9db3-4099-b06e-3199437682ef","Type":"ContainerStarted","Data":"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574"} Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.184448 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d96cdbd-9db3-4099-b06e-3199437682ef","Type":"ContainerStarted","Data":"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6"} Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.202225 4876 generic.go:334] "Generic (PLEG): container finished" podID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerID="148e33290be9b0f22e06f90a2510aab995fd0af4f7d3acd9c95fa7efa28dd22b" exitCode=0 Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.203681 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" event={"ID":"43cdaeee-26ec-47e3-88e9-eda01a1dde0b","Type":"ContainerDied","Data":"148e33290be9b0f22e06f90a2510aab995fd0af4f7d3acd9c95fa7efa28dd22b"} Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.234631 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.234393633 podStartE2EDuration="2.234393633s" podCreationTimestamp="2026-03-13 08:01:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:11.222774698 +0000 UTC m=+1330.893553700" watchObservedRunningTime="2026-03-13 08:01:11.234393633 +0000 UTC m=+1330.905172605" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.268587 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.462420 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.490480 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-nb\") pod \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.490650 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-config\") pod \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.490713 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l95pl\" (UniqueName: \"kubernetes.io/projected/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-kube-api-access-l95pl\") pod \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.490762 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-svc\") pod \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.490952 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-swift-storage-0\") pod \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.490980 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-sb\") pod \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\" (UID: \"43cdaeee-26ec-47e3-88e9-eda01a1dde0b\") " Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.520057 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-kube-api-access-l95pl" (OuterVolumeSpecName: "kube-api-access-l95pl") pod "43cdaeee-26ec-47e3-88e9-eda01a1dde0b" (UID: "43cdaeee-26ec-47e3-88e9-eda01a1dde0b"). InnerVolumeSpecName "kube-api-access-l95pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.570562 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.571141 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.601698 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "43cdaeee-26ec-47e3-88e9-eda01a1dde0b" (UID: "43cdaeee-26ec-47e3-88e9-eda01a1dde0b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.615350 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.615471 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l95pl\" (UniqueName: \"kubernetes.io/projected/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-kube-api-access-l95pl\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.640219 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43cdaeee-26ec-47e3-88e9-eda01a1dde0b" (UID: "43cdaeee-26ec-47e3-88e9-eda01a1dde0b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.647356 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-config" (OuterVolumeSpecName: "config") pod "43cdaeee-26ec-47e3-88e9-eda01a1dde0b" (UID: "43cdaeee-26ec-47e3-88e9-eda01a1dde0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.669816 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "43cdaeee-26ec-47e3-88e9-eda01a1dde0b" (UID: "43cdaeee-26ec-47e3-88e9-eda01a1dde0b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.674789 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "43cdaeee-26ec-47e3-88e9-eda01a1dde0b" (UID: "43cdaeee-26ec-47e3-88e9-eda01a1dde0b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.717717 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.717763 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.717778 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-config\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:11 crc kubenswrapper[4876]: I0313 08:01:11.717788 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43cdaeee-26ec-47e3-88e9-eda01a1dde0b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:12 crc kubenswrapper[4876]: I0313 08:01:12.218546 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" Mar 13 08:01:12 crc kubenswrapper[4876]: I0313 08:01:12.221803 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" event={"ID":"43cdaeee-26ec-47e3-88e9-eda01a1dde0b","Type":"ContainerDied","Data":"c43e0ee356b7f1d208345facb16858c6a34eca75fd97df3f51ac8a1ab2da29e6"} Mar 13 08:01:12 crc kubenswrapper[4876]: I0313 08:01:12.221869 4876 scope.go:117] "RemoveContainer" containerID="148e33290be9b0f22e06f90a2510aab995fd0af4f7d3acd9c95fa7efa28dd22b" Mar 13 08:01:12 crc kubenswrapper[4876]: I0313 08:01:12.258177 4876 scope.go:117] "RemoveContainer" containerID="8b0b01e6fbcbef1c8dbf931b5f260df83af53d917f840ae99c4e7e5295d70609" Mar 13 08:01:12 crc kubenswrapper[4876]: I0313 08:01:12.373055 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-7rqrl"] Mar 13 08:01:12 crc kubenswrapper[4876]: I0313 08:01:12.381783 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-7rqrl"] Mar 13 08:01:13 crc kubenswrapper[4876]: I0313 08:01:13.057010 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" path="/var/lib/kubelet/pods/43cdaeee-26ec-47e3-88e9-eda01a1dde0b/volumes" Mar 13 08:01:13 crc kubenswrapper[4876]: I0313 08:01:13.231016 4876 generic.go:334] "Generic (PLEG): container finished" podID="39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" containerID="58ded0c1d939e2e1aaaa3bcfa7a860776ba705aba34f9551cc684c9b8cdaabd6" exitCode=0 Mar 13 08:01:13 crc kubenswrapper[4876]: I0313 08:01:13.231129 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lrrb8" event={"ID":"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4","Type":"ContainerDied","Data":"58ded0c1d939e2e1aaaa3bcfa7a860776ba705aba34f9551cc684c9b8cdaabd6"} Mar 13 08:01:13 crc kubenswrapper[4876]: I0313 08:01:13.234984 4876 generic.go:334] "Generic (PLEG): container finished" podID="812ab290-71fc-4db3-94af-6daffce9da81" containerID="d3e8eecbdc105c24b74b8b8b513b6288e98ca1cd156c41fa36e38a8cc9ef88d4" exitCode=0 Mar 13 08:01:13 crc kubenswrapper[4876]: I0313 08:01:13.235078 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5st8q" event={"ID":"812ab290-71fc-4db3-94af-6daffce9da81","Type":"ContainerDied","Data":"d3e8eecbdc105c24b74b8b8b513b6288e98ca1cd156c41fa36e38a8cc9ef88d4"} Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.549583 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.550173 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.737181 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.753492 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.892042 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzlvc\" (UniqueName: \"kubernetes.io/projected/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-kube-api-access-mzlvc\") pod \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.892140 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-scripts\") pod \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.892332 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-config-data\") pod \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.892453 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-config-data\") pod \"812ab290-71fc-4db3-94af-6daffce9da81\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.892519 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-combined-ca-bundle\") pod \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\" (UID: \"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4\") " Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.892552 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-scripts\") pod \"812ab290-71fc-4db3-94af-6daffce9da81\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.892665 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-combined-ca-bundle\") pod \"812ab290-71fc-4db3-94af-6daffce9da81\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.892788 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8fsb\" (UniqueName: \"kubernetes.io/projected/812ab290-71fc-4db3-94af-6daffce9da81-kube-api-access-d8fsb\") pod \"812ab290-71fc-4db3-94af-6daffce9da81\" (UID: \"812ab290-71fc-4db3-94af-6daffce9da81\") " Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.899452 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-scripts" (OuterVolumeSpecName: "scripts") pod "812ab290-71fc-4db3-94af-6daffce9da81" (UID: "812ab290-71fc-4db3-94af-6daffce9da81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.900136 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-scripts" (OuterVolumeSpecName: "scripts") pod "39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" (UID: "39f93996-aba4-43e8-aa87-2e0a2cd9c1d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.900790 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-kube-api-access-mzlvc" (OuterVolumeSpecName: "kube-api-access-mzlvc") pod "39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" (UID: "39f93996-aba4-43e8-aa87-2e0a2cd9c1d4"). InnerVolumeSpecName "kube-api-access-mzlvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.906562 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/812ab290-71fc-4db3-94af-6daffce9da81-kube-api-access-d8fsb" (OuterVolumeSpecName: "kube-api-access-d8fsb") pod "812ab290-71fc-4db3-94af-6daffce9da81" (UID: "812ab290-71fc-4db3-94af-6daffce9da81"). InnerVolumeSpecName "kube-api-access-d8fsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.924626 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" (UID: "39f93996-aba4-43e8-aa87-2e0a2cd9c1d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.926342 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "812ab290-71fc-4db3-94af-6daffce9da81" (UID: "812ab290-71fc-4db3-94af-6daffce9da81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.926683 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-config-data" (OuterVolumeSpecName: "config-data") pod "812ab290-71fc-4db3-94af-6daffce9da81" (UID: "812ab290-71fc-4db3-94af-6daffce9da81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.929799 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-config-data" (OuterVolumeSpecName: "config-data") pod "39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" (UID: "39f93996-aba4-43e8-aa87-2e0a2cd9c1d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.995839 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8fsb\" (UniqueName: \"kubernetes.io/projected/812ab290-71fc-4db3-94af-6daffce9da81-kube-api-access-d8fsb\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.995893 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzlvc\" (UniqueName: \"kubernetes.io/projected/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-kube-api-access-mzlvc\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.995914 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.995928 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.995941 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.995953 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.995964 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:14 crc kubenswrapper[4876]: I0313 08:01:14.995976 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/812ab290-71fc-4db3-94af-6daffce9da81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.289444 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lrrb8" event={"ID":"39f93996-aba4-43e8-aa87-2e0a2cd9c1d4","Type":"ContainerDied","Data":"7a9b1d38cbd3bce6e59f86fcda313ebce52cde04fe26f8b6444742a3d2393f17"} Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.290347 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a9b1d38cbd3bce6e59f86fcda313ebce52cde04fe26f8b6444742a3d2393f17" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.289599 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lrrb8" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.308983 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5st8q" event={"ID":"812ab290-71fc-4db3-94af-6daffce9da81","Type":"ContainerDied","Data":"18a152bbeec98549cdbd6dcce0ebbdae51fa0de518d792585b479fcce0345939"} Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.309293 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18a152bbeec98549cdbd6dcce0ebbdae51fa0de518d792585b479fcce0345939" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.309434 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5st8q" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.360665 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 08:01:15 crc kubenswrapper[4876]: E0313 08:01:15.361298 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerName="dnsmasq-dns" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.361322 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerName="dnsmasq-dns" Mar 13 08:01:15 crc kubenswrapper[4876]: E0313 08:01:15.361358 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" containerName="nova-manage" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.361367 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" containerName="nova-manage" Mar 13 08:01:15 crc kubenswrapper[4876]: E0313 08:01:15.361388 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerName="init" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.361394 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerName="init" Mar 13 08:01:15 crc kubenswrapper[4876]: E0313 08:01:15.361409 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812ab290-71fc-4db3-94af-6daffce9da81" containerName="nova-cell1-conductor-db-sync" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.361415 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="812ab290-71fc-4db3-94af-6daffce9da81" containerName="nova-cell1-conductor-db-sync" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.361615 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="812ab290-71fc-4db3-94af-6daffce9da81" containerName="nova-cell1-conductor-db-sync" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.361633 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" containerName="nova-manage" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.361645 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerName="dnsmasq-dns" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.362467 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.365361 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.375842 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.465209 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.465545 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-log" containerID="cri-o://e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6" gracePeriod=30 Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.465984 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-api" containerID="cri-o://c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a" gracePeriod=30 Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.478662 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.478987 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="92a20643-48f7-4b68-a270-358ee507a8cf" containerName="nova-scheduler-scheduler" containerID="cri-o://3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0" gracePeriod=30 Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.509700 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.510034 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.510137 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c7k8\" (UniqueName: \"kubernetes.io/projected/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-kube-api-access-4c7k8\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.544141 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.544588 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerName="nova-metadata-log" containerID="cri-o://c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6" gracePeriod=30 Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.544713 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerName="nova-metadata-metadata" containerID="cri-o://023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574" gracePeriod=30 Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.612253 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.613002 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.613106 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c7k8\" (UniqueName: \"kubernetes.io/projected/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-kube-api-access-4c7k8\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.618137 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.618890 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.637470 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c7k8\" (UniqueName: \"kubernetes.io/projected/1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83-kube-api-access-4c7k8\") pod \"nova-cell1-conductor-0\" (UID: \"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83\") " pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:15 crc kubenswrapper[4876]: I0313 08:01:15.690831 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.191027 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9776ccc5-7rqrl" podUID="43cdaeee-26ec-47e3-88e9-eda01a1dde0b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.166:5353: i/o timeout" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.199522 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.325014 4876 generic.go:334] "Generic (PLEG): container finished" podID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerID="e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6" exitCode=143 Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.325066 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdd5a59b-fa90-4d42-96bf-bc318917490f","Type":"ContainerDied","Data":"e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6"} Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.328389 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d96cdbd-9db3-4099-b06e-3199437682ef-logs\") pod \"9d96cdbd-9db3-4099-b06e-3199437682ef\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.328491 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-config-data\") pod \"9d96cdbd-9db3-4099-b06e-3199437682ef\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.328542 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x948s\" (UniqueName: \"kubernetes.io/projected/9d96cdbd-9db3-4099-b06e-3199437682ef-kube-api-access-x948s\") pod \"9d96cdbd-9db3-4099-b06e-3199437682ef\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.328611 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-combined-ca-bundle\") pod \"9d96cdbd-9db3-4099-b06e-3199437682ef\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.328678 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-nova-metadata-tls-certs\") pod \"9d96cdbd-9db3-4099-b06e-3199437682ef\" (UID: \"9d96cdbd-9db3-4099-b06e-3199437682ef\") " Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.328835 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d96cdbd-9db3-4099-b06e-3199437682ef-logs" (OuterVolumeSpecName: "logs") pod "9d96cdbd-9db3-4099-b06e-3199437682ef" (UID: "9d96cdbd-9db3-4099-b06e-3199437682ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.329281 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d96cdbd-9db3-4099-b06e-3199437682ef-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:16 crc kubenswrapper[4876]: W0313 08:01:16.332028 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f7187f5_a8b9_47ba_99bd_90a6a1cf5d83.slice/crio-067e88c1befc64a7148663d082c1659d09312b6cede36de2b44b0b50f7a0e735 WatchSource:0}: Error finding container 067e88c1befc64a7148663d082c1659d09312b6cede36de2b44b0b50f7a0e735: Status 404 returned error can't find the container with id 067e88c1befc64a7148663d082c1659d09312b6cede36de2b44b0b50f7a0e735 Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.334366 4876 generic.go:334] "Generic (PLEG): container finished" podID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerID="023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574" exitCode=0 Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.334399 4876 generic.go:334] "Generic (PLEG): container finished" podID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerID="c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6" exitCode=143 Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.334419 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d96cdbd-9db3-4099-b06e-3199437682ef","Type":"ContainerDied","Data":"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574"} Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.334450 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d96cdbd-9db3-4099-b06e-3199437682ef","Type":"ContainerDied","Data":"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6"} Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.334463 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9d96cdbd-9db3-4099-b06e-3199437682ef","Type":"ContainerDied","Data":"f09cf9953fe8b802aa40612e9ab01eb50f59f0fa26f41364f1db0105c7fcb368"} Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.334480 4876 scope.go:117] "RemoveContainer" containerID="023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.334485 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.336734 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d96cdbd-9db3-4099-b06e-3199437682ef-kube-api-access-x948s" (OuterVolumeSpecName: "kube-api-access-x948s") pod "9d96cdbd-9db3-4099-b06e-3199437682ef" (UID: "9d96cdbd-9db3-4099-b06e-3199437682ef"). InnerVolumeSpecName "kube-api-access-x948s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.341791 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.364545 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d96cdbd-9db3-4099-b06e-3199437682ef" (UID: "9d96cdbd-9db3-4099-b06e-3199437682ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.368602 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-config-data" (OuterVolumeSpecName: "config-data") pod "9d96cdbd-9db3-4099-b06e-3199437682ef" (UID: "9d96cdbd-9db3-4099-b06e-3199437682ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.384134 4876 scope.go:117] "RemoveContainer" containerID="c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.400544 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9d96cdbd-9db3-4099-b06e-3199437682ef" (UID: "9d96cdbd-9db3-4099-b06e-3199437682ef"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.427218 4876 scope.go:117] "RemoveContainer" containerID="023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574" Mar 13 08:01:16 crc kubenswrapper[4876]: E0313 08:01:16.427970 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574\": container with ID starting with 023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574 not found: ID does not exist" containerID="023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.428039 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574"} err="failed to get container status \"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574\": rpc error: code = NotFound desc = could not find container \"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574\": container with ID starting with 023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574 not found: ID does not exist" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.428078 4876 scope.go:117] "RemoveContainer" containerID="c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6" Mar 13 08:01:16 crc kubenswrapper[4876]: E0313 08:01:16.428749 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6\": container with ID starting with c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6 not found: ID does not exist" containerID="c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.428797 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6"} err="failed to get container status \"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6\": rpc error: code = NotFound desc = could not find container \"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6\": container with ID starting with c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6 not found: ID does not exist" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.428829 4876 scope.go:117] "RemoveContainer" containerID="023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.429323 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574"} err="failed to get container status \"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574\": rpc error: code = NotFound desc = could not find container \"023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574\": container with ID starting with 023fe8c893a2e37e4ec9641dfd410a9cb62914074d8bb644267eff11e2d95574 not found: ID does not exist" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.429353 4876 scope.go:117] "RemoveContainer" containerID="c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.429725 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6"} err="failed to get container status \"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6\": rpc error: code = NotFound desc = could not find container \"c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6\": container with ID starting with c4743058fdc74b4338328532ea53d37f64256eb7e99d298081c720d2547898f6 not found: ID does not exist" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.430831 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.430861 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x948s\" (UniqueName: \"kubernetes.io/projected/9d96cdbd-9db3-4099-b06e-3199437682ef-kube-api-access-x948s\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.430875 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.430886 4876 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d96cdbd-9db3-4099-b06e-3199437682ef-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.726095 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.738462 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.754302 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:16 crc kubenswrapper[4876]: E0313 08:01:16.755645 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerName="nova-metadata-log" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.755677 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerName="nova-metadata-log" Mar 13 08:01:16 crc kubenswrapper[4876]: E0313 08:01:16.755717 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerName="nova-metadata-metadata" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.755729 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerName="nova-metadata-metadata" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.755992 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerName="nova-metadata-metadata" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.756018 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" containerName="nova-metadata-log" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.757376 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.761408 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.761429 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.784687 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.943358 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.943509 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.943548 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7qfz\" (UniqueName: \"kubernetes.io/projected/020dff52-bda7-470d-8e33-7faf45341286-kube-api-access-n7qfz\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.943628 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-config-data\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:16 crc kubenswrapper[4876]: I0313 08:01:16.943683 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/020dff52-bda7-470d-8e33-7faf45341286-logs\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.045646 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.045733 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.045784 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7qfz\" (UniqueName: \"kubernetes.io/projected/020dff52-bda7-470d-8e33-7faf45341286-kube-api-access-n7qfz\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.045860 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-config-data\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.045893 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/020dff52-bda7-470d-8e33-7faf45341286-logs\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.046615 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/020dff52-bda7-470d-8e33-7faf45341286-logs\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.049600 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d96cdbd-9db3-4099-b06e-3199437682ef" path="/var/lib/kubelet/pods/9d96cdbd-9db3-4099-b06e-3199437682ef/volumes" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.052991 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.052994 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.060034 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-config-data\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.069999 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7qfz\" (UniqueName: \"kubernetes.io/projected/020dff52-bda7-470d-8e33-7faf45341286-kube-api-access-n7qfz\") pod \"nova-metadata-0\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.079023 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.353089 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83","Type":"ContainerStarted","Data":"f79c05078c880c51d12337b3c3cf06216b2c2b12a73cb9b2fa02a3a84bcdfc7b"} Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.353150 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83","Type":"ContainerStarted","Data":"067e88c1befc64a7148663d082c1659d09312b6cede36de2b44b0b50f7a0e735"} Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.353425 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.385934 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.385907642 podStartE2EDuration="2.385907642s" podCreationTimestamp="2026-03-13 08:01:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:17.375280006 +0000 UTC m=+1337.046058988" watchObservedRunningTime="2026-03-13 08:01:17.385907642 +0000 UTC m=+1337.056686624" Mar 13 08:01:17 crc kubenswrapper[4876]: I0313 08:01:17.598768 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:01:18 crc kubenswrapper[4876]: I0313 08:01:18.330050 4876 scope.go:117] "RemoveContainer" containerID="4b928f88852dc04e2f1b3cad2b15db0da8b97c2bc616434bde6ec1839f321c5b" Mar 13 08:01:18 crc kubenswrapper[4876]: I0313 08:01:18.370483 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"020dff52-bda7-470d-8e33-7faf45341286","Type":"ContainerStarted","Data":"b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a"} Mar 13 08:01:18 crc kubenswrapper[4876]: I0313 08:01:18.370550 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"020dff52-bda7-470d-8e33-7faf45341286","Type":"ContainerStarted","Data":"fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c"} Mar 13 08:01:18 crc kubenswrapper[4876]: I0313 08:01:18.370566 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"020dff52-bda7-470d-8e33-7faf45341286","Type":"ContainerStarted","Data":"8b0d8fe6bf7aa4c2451af6c7de00a3a1afeec4d2e95b19ec2577a4f081f2cb19"} Mar 13 08:01:18 crc kubenswrapper[4876]: I0313 08:01:18.404940 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.404906095 podStartE2EDuration="2.404906095s" podCreationTimestamp="2026-03-13 08:01:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:18.396358776 +0000 UTC m=+1338.067137788" watchObservedRunningTime="2026-03-13 08:01:18.404906095 +0000 UTC m=+1338.075685087" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.207259 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.304136 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-config-data\") pod \"cdd5a59b-fa90-4d42-96bf-bc318917490f\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.304310 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvvq2\" (UniqueName: \"kubernetes.io/projected/cdd5a59b-fa90-4d42-96bf-bc318917490f-kube-api-access-cvvq2\") pod \"cdd5a59b-fa90-4d42-96bf-bc318917490f\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.304620 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd5a59b-fa90-4d42-96bf-bc318917490f-logs\") pod \"cdd5a59b-fa90-4d42-96bf-bc318917490f\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.304693 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-combined-ca-bundle\") pod \"cdd5a59b-fa90-4d42-96bf-bc318917490f\" (UID: \"cdd5a59b-fa90-4d42-96bf-bc318917490f\") " Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.305916 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdd5a59b-fa90-4d42-96bf-bc318917490f-logs" (OuterVolumeSpecName: "logs") pod "cdd5a59b-fa90-4d42-96bf-bc318917490f" (UID: "cdd5a59b-fa90-4d42-96bf-bc318917490f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.311965 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdd5a59b-fa90-4d42-96bf-bc318917490f-kube-api-access-cvvq2" (OuterVolumeSpecName: "kube-api-access-cvvq2") pod "cdd5a59b-fa90-4d42-96bf-bc318917490f" (UID: "cdd5a59b-fa90-4d42-96bf-bc318917490f"). InnerVolumeSpecName "kube-api-access-cvvq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.340385 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdd5a59b-fa90-4d42-96bf-bc318917490f" (UID: "cdd5a59b-fa90-4d42-96bf-bc318917490f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.344953 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-config-data" (OuterVolumeSpecName: "config-data") pod "cdd5a59b-fa90-4d42-96bf-bc318917490f" (UID: "cdd5a59b-fa90-4d42-96bf-bc318917490f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.386921 4876 generic.go:334] "Generic (PLEG): container finished" podID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerID="c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a" exitCode=0 Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.388499 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.392707 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdd5a59b-fa90-4d42-96bf-bc318917490f","Type":"ContainerDied","Data":"c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a"} Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.392795 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdd5a59b-fa90-4d42-96bf-bc318917490f","Type":"ContainerDied","Data":"cbd4317f4dd6b931c50ea7a70dda3895078781f344069a51b985e45b6ae51861"} Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.392837 4876 scope.go:117] "RemoveContainer" containerID="c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.406743 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd5a59b-fa90-4d42-96bf-bc318917490f-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.406790 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.406800 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd5a59b-fa90-4d42-96bf-bc318917490f-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.406811 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvvq2\" (UniqueName: \"kubernetes.io/projected/cdd5a59b-fa90-4d42-96bf-bc318917490f-kube-api-access-cvvq2\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.500326 4876 scope.go:117] "RemoveContainer" containerID="e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.506864 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.524559 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.534956 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:19 crc kubenswrapper[4876]: E0313 08:01:19.539870 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-log" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.539906 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-log" Mar 13 08:01:19 crc kubenswrapper[4876]: E0313 08:01:19.539966 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-api" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.539972 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-api" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.540234 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-log" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.540277 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" containerName="nova-api-api" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.541340 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.545748 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.550221 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.554412 4876 scope.go:117] "RemoveContainer" containerID="c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a" Mar 13 08:01:19 crc kubenswrapper[4876]: E0313 08:01:19.555782 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a\": container with ID starting with c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a not found: ID does not exist" containerID="c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.555836 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a"} err="failed to get container status \"c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a\": rpc error: code = NotFound desc = could not find container \"c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a\": container with ID starting with c48ebbe04403d350ab83023575488c025f0fe5e840686408bf4addb03ec4920a not found: ID does not exist" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.555871 4876 scope.go:117] "RemoveContainer" containerID="e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6" Mar 13 08:01:19 crc kubenswrapper[4876]: E0313 08:01:19.559541 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6\": container with ID starting with e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6 not found: ID does not exist" containerID="e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.559636 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6"} err="failed to get container status \"e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6\": rpc error: code = NotFound desc = could not find container \"e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6\": container with ID starting with e628df1b01f230710ef53ba7696a991751b0ca3955b0f4fa0b972c283585c7c6 not found: ID does not exist" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.701529 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.715558 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ca6048-fb47-45bb-9bc0-a441d684c798-logs\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.715649 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72w8c\" (UniqueName: \"kubernetes.io/projected/23ca6048-fb47-45bb-9bc0-a441d684c798-kube-api-access-72w8c\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.715718 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-config-data\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.715784 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.817636 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-config-data\") pod \"92a20643-48f7-4b68-a270-358ee507a8cf\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.818028 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b297r\" (UniqueName: \"kubernetes.io/projected/92a20643-48f7-4b68-a270-358ee507a8cf-kube-api-access-b297r\") pod \"92a20643-48f7-4b68-a270-358ee507a8cf\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.818055 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-combined-ca-bundle\") pod \"92a20643-48f7-4b68-a270-358ee507a8cf\" (UID: \"92a20643-48f7-4b68-a270-358ee507a8cf\") " Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.818352 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ca6048-fb47-45bb-9bc0-a441d684c798-logs\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.818410 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72w8c\" (UniqueName: \"kubernetes.io/projected/23ca6048-fb47-45bb-9bc0-a441d684c798-kube-api-access-72w8c\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.818438 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-config-data\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.818489 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.819207 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ca6048-fb47-45bb-9bc0-a441d684c798-logs\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.823825 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.825358 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a20643-48f7-4b68-a270-358ee507a8cf-kube-api-access-b297r" (OuterVolumeSpecName: "kube-api-access-b297r") pod "92a20643-48f7-4b68-a270-358ee507a8cf" (UID: "92a20643-48f7-4b68-a270-358ee507a8cf"). InnerVolumeSpecName "kube-api-access-b297r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.828099 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-config-data\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.837329 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72w8c\" (UniqueName: \"kubernetes.io/projected/23ca6048-fb47-45bb-9bc0-a441d684c798-kube-api-access-72w8c\") pod \"nova-api-0\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.849747 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-config-data" (OuterVolumeSpecName: "config-data") pod "92a20643-48f7-4b68-a270-358ee507a8cf" (UID: "92a20643-48f7-4b68-a270-358ee507a8cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.857932 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92a20643-48f7-4b68-a270-358ee507a8cf" (UID: "92a20643-48f7-4b68-a270-358ee507a8cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.889945 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.921023 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.921076 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b297r\" (UniqueName: \"kubernetes.io/projected/92a20643-48f7-4b68-a270-358ee507a8cf-kube-api-access-b297r\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:19 crc kubenswrapper[4876]: I0313 08:01:19.921091 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92a20643-48f7-4b68-a270-358ee507a8cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.359415 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.416177 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ca6048-fb47-45bb-9bc0-a441d684c798","Type":"ContainerStarted","Data":"84192caa1c00be3c5ae6844baf19e16845969b93f2b04a9a4f35b56ce26ae49a"} Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.421583 4876 generic.go:334] "Generic (PLEG): container finished" podID="92a20643-48f7-4b68-a270-358ee507a8cf" containerID="3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0" exitCode=0 Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.421945 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.422176 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92a20643-48f7-4b68-a270-358ee507a8cf","Type":"ContainerDied","Data":"3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0"} Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.422332 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92a20643-48f7-4b68-a270-358ee507a8cf","Type":"ContainerDied","Data":"dce0b4f8bf70050cda0d0905fb110097e62c2919f3d25296d8fb8fb407230f53"} Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.427485 4876 scope.go:117] "RemoveContainer" containerID="3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.460942 4876 scope.go:117] "RemoveContainer" containerID="3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0" Mar 13 08:01:20 crc kubenswrapper[4876]: E0313 08:01:20.465497 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0\": container with ID starting with 3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0 not found: ID does not exist" containerID="3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.465540 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0"} err="failed to get container status \"3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0\": rpc error: code = NotFound desc = could not find container \"3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0\": container with ID starting with 3c1a939027b8e1377056bcd5d31b293491ea6f076371dacc2c4beaed8a856ae0 not found: ID does not exist" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.481457 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.499462 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.520540 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:20 crc kubenswrapper[4876]: E0313 08:01:20.522563 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a20643-48f7-4b68-a270-358ee507a8cf" containerName="nova-scheduler-scheduler" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.522595 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a20643-48f7-4b68-a270-358ee507a8cf" containerName="nova-scheduler-scheduler" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.522927 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a20643-48f7-4b68-a270-358ee507a8cf" containerName="nova-scheduler-scheduler" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.523848 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.528948 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.547513 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.647470 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-config-data\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.647625 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqf8z\" (UniqueName: \"kubernetes.io/projected/1da86b72-7254-41cf-a337-c5f309865f27-kube-api-access-hqf8z\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.647999 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.751024 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.751158 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-config-data\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.751290 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqf8z\" (UniqueName: \"kubernetes.io/projected/1da86b72-7254-41cf-a337-c5f309865f27-kube-api-access-hqf8z\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.758260 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.760014 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-config-data\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.772831 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqf8z\" (UniqueName: \"kubernetes.io/projected/1da86b72-7254-41cf-a337-c5f309865f27-kube-api-access-hqf8z\") pod \"nova-scheduler-0\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " pod="openstack/nova-scheduler-0" Mar 13 08:01:20 crc kubenswrapper[4876]: I0313 08:01:20.853301 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:01:21 crc kubenswrapper[4876]: I0313 08:01:21.052708 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a20643-48f7-4b68-a270-358ee507a8cf" path="/var/lib/kubelet/pods/92a20643-48f7-4b68-a270-358ee507a8cf/volumes" Mar 13 08:01:21 crc kubenswrapper[4876]: I0313 08:01:21.054213 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdd5a59b-fa90-4d42-96bf-bc318917490f" path="/var/lib/kubelet/pods/cdd5a59b-fa90-4d42-96bf-bc318917490f/volumes" Mar 13 08:01:21 crc kubenswrapper[4876]: W0313 08:01:21.359375 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1da86b72_7254_41cf_a337_c5f309865f27.slice/crio-ba8c0e33882e94f1dc207256a2dc91bf682746f1f5b21f675a509f90ee09d63f WatchSource:0}: Error finding container ba8c0e33882e94f1dc207256a2dc91bf682746f1f5b21f675a509f90ee09d63f: Status 404 returned error can't find the container with id ba8c0e33882e94f1dc207256a2dc91bf682746f1f5b21f675a509f90ee09d63f Mar 13 08:01:21 crc kubenswrapper[4876]: I0313 08:01:21.361630 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:01:21 crc kubenswrapper[4876]: I0313 08:01:21.443830 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ca6048-fb47-45bb-9bc0-a441d684c798","Type":"ContainerStarted","Data":"0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676"} Mar 13 08:01:21 crc kubenswrapper[4876]: I0313 08:01:21.443936 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ca6048-fb47-45bb-9bc0-a441d684c798","Type":"ContainerStarted","Data":"afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847"} Mar 13 08:01:21 crc kubenswrapper[4876]: I0313 08:01:21.449145 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1da86b72-7254-41cf-a337-c5f309865f27","Type":"ContainerStarted","Data":"ba8c0e33882e94f1dc207256a2dc91bf682746f1f5b21f675a509f90ee09d63f"} Mar 13 08:01:21 crc kubenswrapper[4876]: I0313 08:01:21.478523 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.478501154 podStartE2EDuration="2.478501154s" podCreationTimestamp="2026-03-13 08:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:21.464490792 +0000 UTC m=+1341.135269774" watchObservedRunningTime="2026-03-13 08:01:21.478501154 +0000 UTC m=+1341.149280136" Mar 13 08:01:22 crc kubenswrapper[4876]: I0313 08:01:22.080137 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 08:01:22 crc kubenswrapper[4876]: I0313 08:01:22.080715 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 08:01:22 crc kubenswrapper[4876]: I0313 08:01:22.462176 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1da86b72-7254-41cf-a337-c5f309865f27","Type":"ContainerStarted","Data":"db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541"} Mar 13 08:01:22 crc kubenswrapper[4876]: I0313 08:01:22.494186 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.494137712 podStartE2EDuration="2.494137712s" podCreationTimestamp="2026-03-13 08:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:22.490881261 +0000 UTC m=+1342.161660243" watchObservedRunningTime="2026-03-13 08:01:22.494137712 +0000 UTC m=+1342.164916684" Mar 13 08:01:25 crc kubenswrapper[4876]: I0313 08:01:25.733213 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 08:01:25 crc kubenswrapper[4876]: I0313 08:01:25.854172 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 08:01:26 crc kubenswrapper[4876]: I0313 08:01:26.035180 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 08:01:27 crc kubenswrapper[4876]: I0313 08:01:27.079970 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 08:01:27 crc kubenswrapper[4876]: I0313 08:01:27.080447 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 08:01:28 crc kubenswrapper[4876]: I0313 08:01:28.093504 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 08:01:28 crc kubenswrapper[4876]: I0313 08:01:28.093561 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 08:01:29 crc kubenswrapper[4876]: I0313 08:01:29.890899 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 08:01:29 crc kubenswrapper[4876]: I0313 08:01:29.891415 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 08:01:29 crc kubenswrapper[4876]: I0313 08:01:29.912699 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 08:01:29 crc kubenswrapper[4876]: I0313 08:01:29.913004 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="7a6b118f-c518-4e1e-9f0a-79bb844a304f" containerName="kube-state-metrics" containerID="cri-o://9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345" gracePeriod=30 Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.460346 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.561682 4876 generic.go:334] "Generic (PLEG): container finished" podID="7a6b118f-c518-4e1e-9f0a-79bb844a304f" containerID="9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345" exitCode=2 Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.561735 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7a6b118f-c518-4e1e-9f0a-79bb844a304f","Type":"ContainerDied","Data":"9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345"} Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.561785 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7a6b118f-c518-4e1e-9f0a-79bb844a304f","Type":"ContainerDied","Data":"ed426a1f40e97f2e7e362ae43bc3ee56c7291153e7af4a7ed2a58dbb1db7e12f"} Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.561806 4876 scope.go:117] "RemoveContainer" containerID="9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.562228 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.581620 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69g2n\" (UniqueName: \"kubernetes.io/projected/7a6b118f-c518-4e1e-9f0a-79bb844a304f-kube-api-access-69g2n\") pod \"7a6b118f-c518-4e1e-9f0a-79bb844a304f\" (UID: \"7a6b118f-c518-4e1e-9f0a-79bb844a304f\") " Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.591933 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6b118f-c518-4e1e-9f0a-79bb844a304f-kube-api-access-69g2n" (OuterVolumeSpecName: "kube-api-access-69g2n") pod "7a6b118f-c518-4e1e-9f0a-79bb844a304f" (UID: "7a6b118f-c518-4e1e-9f0a-79bb844a304f"). InnerVolumeSpecName "kube-api-access-69g2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.599953 4876 scope.go:117] "RemoveContainer" containerID="9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345" Mar 13 08:01:30 crc kubenswrapper[4876]: E0313 08:01:30.600657 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345\": container with ID starting with 9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345 not found: ID does not exist" containerID="9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.600764 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345"} err="failed to get container status \"9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345\": rpc error: code = NotFound desc = could not find container \"9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345\": container with ID starting with 9f6cd60c658bc8a125a48bfe008c12e00adb54914457d01fa6c29ba8ea628345 not found: ID does not exist" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.685460 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69g2n\" (UniqueName: \"kubernetes.io/projected/7a6b118f-c518-4e1e-9f0a-79bb844a304f-kube-api-access-69g2n\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.854649 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.903548 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.906519 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.916428 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.927262 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 08:01:30 crc kubenswrapper[4876]: E0313 08:01:30.927822 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6b118f-c518-4e1e-9f0a-79bb844a304f" containerName="kube-state-metrics" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.927844 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6b118f-c518-4e1e-9f0a-79bb844a304f" containerName="kube-state-metrics" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.928104 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6b118f-c518-4e1e-9f0a-79bb844a304f" containerName="kube-state-metrics" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.929032 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.941640 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.941995 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.951062 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.980422 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.200:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 08:01:30 crc kubenswrapper[4876]: I0313 08:01:30.980729 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.200:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.055695 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6b118f-c518-4e1e-9f0a-79bb844a304f" path="/var/lib/kubelet/pods/7a6b118f-c518-4e1e-9f0a-79bb844a304f/volumes" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.108672 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.109403 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.110423 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwrxw\" (UniqueName: \"kubernetes.io/projected/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-api-access-cwrxw\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.111619 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.213740 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.213821 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.213887 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwrxw\" (UniqueName: \"kubernetes.io/projected/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-api-access-cwrxw\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.213941 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.218965 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.227869 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.236452 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff27b563-91dc-4be1-b0b8-98955850d2dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.254020 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwrxw\" (UniqueName: \"kubernetes.io/projected/ff27b563-91dc-4be1-b0b8-98955850d2dd-kube-api-access-cwrxw\") pod \"kube-state-metrics-0\" (UID: \"ff27b563-91dc-4be1-b0b8-98955850d2dd\") " pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.295304 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.616406 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 08:01:31 crc kubenswrapper[4876]: I0313 08:01:31.850823 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.361319 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.362036 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="ceilometer-central-agent" containerID="cri-o://3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491" gracePeriod=30 Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.363122 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="proxy-httpd" containerID="cri-o://ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8" gracePeriod=30 Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.363202 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="sg-core" containerID="cri-o://9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207" gracePeriod=30 Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.363328 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="ceilometer-notification-agent" containerID="cri-o://2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330" gracePeriod=30 Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.594437 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ff27b563-91dc-4be1-b0b8-98955850d2dd","Type":"ContainerStarted","Data":"e4653a8f58be7cfb0f58cf3d606b6efdec0a2e30abe8a6075de8145c5f2ccb53"} Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.597984 4876 generic.go:334] "Generic (PLEG): container finished" podID="3ddfd44e-033d-49c2-b174-c239e223388b" containerID="ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8" exitCode=0 Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.598021 4876 generic.go:334] "Generic (PLEG): container finished" podID="3ddfd44e-033d-49c2-b174-c239e223388b" containerID="9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207" exitCode=2 Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.598035 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerDied","Data":"ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8"} Mar 13 08:01:32 crc kubenswrapper[4876]: I0313 08:01:32.598070 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerDied","Data":"9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207"} Mar 13 08:01:33 crc kubenswrapper[4876]: I0313 08:01:33.610443 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ff27b563-91dc-4be1-b0b8-98955850d2dd","Type":"ContainerStarted","Data":"2338dfbfd423e77f2d64e77f211a204e9c16d87d388683c80ab74ece1d326331"} Mar 13 08:01:33 crc kubenswrapper[4876]: I0313 08:01:33.610563 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 08:01:33 crc kubenswrapper[4876]: I0313 08:01:33.614995 4876 generic.go:334] "Generic (PLEG): container finished" podID="3ddfd44e-033d-49c2-b174-c239e223388b" containerID="3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491" exitCode=0 Mar 13 08:01:33 crc kubenswrapper[4876]: I0313 08:01:33.615054 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerDied","Data":"3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491"} Mar 13 08:01:33 crc kubenswrapper[4876]: I0313 08:01:33.635870 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.179910114 podStartE2EDuration="3.635840663s" podCreationTimestamp="2026-03-13 08:01:30 +0000 UTC" firstStartedPulling="2026-03-13 08:01:31.859207872 +0000 UTC m=+1351.529986854" lastFinishedPulling="2026-03-13 08:01:32.315138421 +0000 UTC m=+1351.985917403" observedRunningTime="2026-03-13 08:01:33.632663554 +0000 UTC m=+1353.303442556" watchObservedRunningTime="2026-03-13 08:01:33.635840663 +0000 UTC m=+1353.306619685" Mar 13 08:01:37 crc kubenswrapper[4876]: I0313 08:01:37.086956 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 08:01:37 crc kubenswrapper[4876]: I0313 08:01:37.104032 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 08:01:37 crc kubenswrapper[4876]: I0313 08:01:37.107359 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 08:01:37 crc kubenswrapper[4876]: I0313 08:01:37.667604 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.241217 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.370956 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frvhr\" (UniqueName: \"kubernetes.io/projected/3ddfd44e-033d-49c2-b174-c239e223388b-kube-api-access-frvhr\") pod \"3ddfd44e-033d-49c2-b174-c239e223388b\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.371089 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-scripts\") pod \"3ddfd44e-033d-49c2-b174-c239e223388b\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.371170 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-config-data\") pod \"3ddfd44e-033d-49c2-b174-c239e223388b\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.371355 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-log-httpd\") pod \"3ddfd44e-033d-49c2-b174-c239e223388b\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.371567 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-combined-ca-bundle\") pod \"3ddfd44e-033d-49c2-b174-c239e223388b\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.371721 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-run-httpd\") pod \"3ddfd44e-033d-49c2-b174-c239e223388b\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.371750 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3ddfd44e-033d-49c2-b174-c239e223388b" (UID: "3ddfd44e-033d-49c2-b174-c239e223388b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.371829 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-sg-core-conf-yaml\") pod \"3ddfd44e-033d-49c2-b174-c239e223388b\" (UID: \"3ddfd44e-033d-49c2-b174-c239e223388b\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.372203 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3ddfd44e-033d-49c2-b174-c239e223388b" (UID: "3ddfd44e-033d-49c2-b174-c239e223388b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.373078 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.373101 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ddfd44e-033d-49c2-b174-c239e223388b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.387693 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ddfd44e-033d-49c2-b174-c239e223388b-kube-api-access-frvhr" (OuterVolumeSpecName: "kube-api-access-frvhr") pod "3ddfd44e-033d-49c2-b174-c239e223388b" (UID: "3ddfd44e-033d-49c2-b174-c239e223388b"). InnerVolumeSpecName "kube-api-access-frvhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.392680 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-scripts" (OuterVolumeSpecName: "scripts") pod "3ddfd44e-033d-49c2-b174-c239e223388b" (UID: "3ddfd44e-033d-49c2-b174-c239e223388b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.442421 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3ddfd44e-033d-49c2-b174-c239e223388b" (UID: "3ddfd44e-033d-49c2-b174-c239e223388b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.447714 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.475650 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frvhr\" (UniqueName: \"kubernetes.io/projected/3ddfd44e-033d-49c2-b174-c239e223388b-kube-api-access-frvhr\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.475693 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.475733 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.483587 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ddfd44e-033d-49c2-b174-c239e223388b" (UID: "3ddfd44e-033d-49c2-b174-c239e223388b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.528861 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-config-data" (OuterVolumeSpecName: "config-data") pod "3ddfd44e-033d-49c2-b174-c239e223388b" (UID: "3ddfd44e-033d-49c2-b174-c239e223388b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.577523 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-combined-ca-bundle\") pod \"b304c04a-a07f-487d-8b8e-7379fa339016\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.577751 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-config-data\") pod \"b304c04a-a07f-487d-8b8e-7379fa339016\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.577788 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwfwg\" (UniqueName: \"kubernetes.io/projected/b304c04a-a07f-487d-8b8e-7379fa339016-kube-api-access-hwfwg\") pod \"b304c04a-a07f-487d-8b8e-7379fa339016\" (UID: \"b304c04a-a07f-487d-8b8e-7379fa339016\") " Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.579338 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.579366 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddfd44e-033d-49c2-b174-c239e223388b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.583630 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b304c04a-a07f-487d-8b8e-7379fa339016-kube-api-access-hwfwg" (OuterVolumeSpecName: "kube-api-access-hwfwg") pod "b304c04a-a07f-487d-8b8e-7379fa339016" (UID: "b304c04a-a07f-487d-8b8e-7379fa339016"). InnerVolumeSpecName "kube-api-access-hwfwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.608361 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b304c04a-a07f-487d-8b8e-7379fa339016" (UID: "b304c04a-a07f-487d-8b8e-7379fa339016"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.614496 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-config-data" (OuterVolumeSpecName: "config-data") pod "b304c04a-a07f-487d-8b8e-7379fa339016" (UID: "b304c04a-a07f-487d-8b8e-7379fa339016"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.674399 4876 generic.go:334] "Generic (PLEG): container finished" podID="b304c04a-a07f-487d-8b8e-7379fa339016" containerID="337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4" exitCode=137 Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.674469 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.674489 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b304c04a-a07f-487d-8b8e-7379fa339016","Type":"ContainerDied","Data":"337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4"} Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.674529 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b304c04a-a07f-487d-8b8e-7379fa339016","Type":"ContainerDied","Data":"cdc52b62b1066b080f9d4f963e23b223a15fee5ab341dea30ca2930433c2bf82"} Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.674587 4876 scope.go:117] "RemoveContainer" containerID="337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.681063 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.681071 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.681192 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b304c04a-a07f-487d-8b8e-7379fa339016-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.681206 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwfwg\" (UniqueName: \"kubernetes.io/projected/b304c04a-a07f-487d-8b8e-7379fa339016-kube-api-access-hwfwg\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.690746 4876 generic.go:334] "Generic (PLEG): container finished" podID="3ddfd44e-033d-49c2-b174-c239e223388b" containerID="2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330" exitCode=0 Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.690946 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerDied","Data":"2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330"} Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.691011 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ddfd44e-033d-49c2-b174-c239e223388b","Type":"ContainerDied","Data":"a66fa78401219287c1244f5322785ee0935873e52a4f522bbc3781f8eb34afe2"} Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.721548 4876 scope.go:117] "RemoveContainer" containerID="337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.722006 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4\": container with ID starting with 337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4 not found: ID does not exist" containerID="337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.722039 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4"} err="failed to get container status \"337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4\": rpc error: code = NotFound desc = could not find container \"337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4\": container with ID starting with 337feb8aad7123c0b97cdeda6ebaf50288337eeb9418f85ec189166c35b052e4 not found: ID does not exist" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.722065 4876 scope.go:117] "RemoveContainer" containerID="ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.757502 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.766718 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.779341 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.780116 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="sg-core" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780139 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="sg-core" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.780183 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="ceilometer-central-agent" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780192 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="ceilometer-central-agent" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.780207 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="ceilometer-notification-agent" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780214 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="ceilometer-notification-agent" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.780257 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="proxy-httpd" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780269 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="proxy-httpd" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.780280 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b304c04a-a07f-487d-8b8e-7379fa339016" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780288 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b304c04a-a07f-487d-8b8e-7379fa339016" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780552 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="ceilometer-central-agent" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780570 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b304c04a-a07f-487d-8b8e-7379fa339016" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780587 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="ceilometer-notification-agent" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780600 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="sg-core" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.780619 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" containerName="proxy-httpd" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.789058 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.795687 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.795906 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.796189 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.797853 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.843349 4876 scope.go:117] "RemoveContainer" containerID="9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.889279 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.889404 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsjgh\" (UniqueName: \"kubernetes.io/projected/9b56d8da-0553-46dd-b2ab-26df263e1885-kube-api-access-vsjgh\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.889699 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.889807 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.889978 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.901433 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.901635 4876 scope.go:117] "RemoveContainer" containerID="2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.914111 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.931385 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.934315 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.934444 4876 scope.go:117] "RemoveContainer" containerID="3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.937651 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.938112 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.938118 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.938552 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.961090 4876 scope.go:117] "RemoveContainer" containerID="ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.961739 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8\": container with ID starting with ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8 not found: ID does not exist" containerID="ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.961783 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8"} err="failed to get container status \"ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8\": rpc error: code = NotFound desc = could not find container \"ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8\": container with ID starting with ad3c7f22fc6442ecdfb8fcbbede88ea28b190b92b5c6071fc961489588e106b8 not found: ID does not exist" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.961810 4876 scope.go:117] "RemoveContainer" containerID="9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.962936 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207\": container with ID starting with 9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207 not found: ID does not exist" containerID="9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.962988 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207"} err="failed to get container status \"9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207\": rpc error: code = NotFound desc = could not find container \"9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207\": container with ID starting with 9488433f39e8e3b6c3419f2677f4185b0c3e6d9ddea3ba7068d30e36fe234207 not found: ID does not exist" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.963024 4876 scope.go:117] "RemoveContainer" containerID="2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.963365 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330\": container with ID starting with 2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330 not found: ID does not exist" containerID="2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.963396 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330"} err="failed to get container status \"2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330\": rpc error: code = NotFound desc = could not find container \"2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330\": container with ID starting with 2a8680113a9ecf84702ce3ddc61020cff3d215cff8240fd07a750b06aa987330 not found: ID does not exist" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.963413 4876 scope.go:117] "RemoveContainer" containerID="3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491" Mar 13 08:01:38 crc kubenswrapper[4876]: E0313 08:01:38.963694 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491\": container with ID starting with 3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491 not found: ID does not exist" containerID="3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.963719 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491"} err="failed to get container status \"3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491\": rpc error: code = NotFound desc = could not find container \"3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491\": container with ID starting with 3a86cff2685a2f3be85369b73f48ac6e80598adaefaebb66cd1e9ce9e48e4491 not found: ID does not exist" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.992228 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsjgh\" (UniqueName: \"kubernetes.io/projected/9b56d8da-0553-46dd-b2ab-26df263e1885-kube-api-access-vsjgh\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.992425 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.992489 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.992566 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.992594 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.997536 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.998547 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.998610 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:38 crc kubenswrapper[4876]: I0313 08:01:38.999694 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b56d8da-0553-46dd-b2ab-26df263e1885-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.010715 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsjgh\" (UniqueName: \"kubernetes.io/projected/9b56d8da-0553-46dd-b2ab-26df263e1885-kube-api-access-vsjgh\") pod \"nova-cell1-novncproxy-0\" (UID: \"9b56d8da-0553-46dd-b2ab-26df263e1885\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.049411 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ddfd44e-033d-49c2-b174-c239e223388b" path="/var/lib/kubelet/pods/3ddfd44e-033d-49c2-b174-c239e223388b/volumes" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.050271 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b304c04a-a07f-487d-8b8e-7379fa339016" path="/var/lib/kubelet/pods/b304c04a-a07f-487d-8b8e-7379fa339016/volumes" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.094910 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-log-httpd\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.095005 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-scripts\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.095048 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-config-data\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.095091 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.095205 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.095259 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-run-httpd\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.095320 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4qvn\" (UniqueName: \"kubernetes.io/projected/f5cdea04-68df-432d-8cd5-9b6f5cba9712-kube-api-access-v4qvn\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.095390 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.150642 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.197427 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-run-httpd\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.197954 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4qvn\" (UniqueName: \"kubernetes.io/projected/f5cdea04-68df-432d-8cd5-9b6f5cba9712-kube-api-access-v4qvn\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.198009 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.198062 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-log-httpd\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.198092 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-scripts\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.198140 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-config-data\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.198199 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.198272 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.198341 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-run-httpd\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.198614 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-log-httpd\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.205014 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.205144 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-scripts\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.205405 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.205619 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.207230 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-config-data\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.217371 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4qvn\" (UniqueName: \"kubernetes.io/projected/f5cdea04-68df-432d-8cd5-9b6f5cba9712-kube-api-access-v4qvn\") pod \"ceilometer-0\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.257104 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.664872 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.710004 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9b56d8da-0553-46dd-b2ab-26df263e1885","Type":"ContainerStarted","Data":"f31306a36905b8aeeec8ed0ea54227f5774b12bb7ca20163ce7f5b35d4caafc3"} Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.800873 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:39 crc kubenswrapper[4876]: W0313 08:01:39.806003 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5cdea04_68df_432d_8cd5_9b6f5cba9712.slice/crio-aead3067c9cbddb025ebc154af04a221278bcc22b0329c38bb1361135ce01520 WatchSource:0}: Error finding container aead3067c9cbddb025ebc154af04a221278bcc22b0329c38bb1361135ce01520: Status 404 returned error can't find the container with id aead3067c9cbddb025ebc154af04a221278bcc22b0329c38bb1361135ce01520 Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.896411 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.897176 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.897294 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 08:01:39 crc kubenswrapper[4876]: I0313 08:01:39.899453 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 08:01:40 crc kubenswrapper[4876]: I0313 08:01:40.724082 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerStarted","Data":"bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf"} Mar 13 08:01:40 crc kubenswrapper[4876]: I0313 08:01:40.724487 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerStarted","Data":"aead3067c9cbddb025ebc154af04a221278bcc22b0329c38bb1361135ce01520"} Mar 13 08:01:40 crc kubenswrapper[4876]: I0313 08:01:40.732496 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9b56d8da-0553-46dd-b2ab-26df263e1885","Type":"ContainerStarted","Data":"b00e008bfe04efbb3159f3f63a9a641c3814cd1c9933eda3f7f390ca8ff387c6"} Mar 13 08:01:40 crc kubenswrapper[4876]: I0313 08:01:40.732863 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 08:01:40 crc kubenswrapper[4876]: I0313 08:01:40.744944 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 08:01:40 crc kubenswrapper[4876]: I0313 08:01:40.775485 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.775448332 podStartE2EDuration="2.775448332s" podCreationTimestamp="2026-03-13 08:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:40.759181257 +0000 UTC m=+1360.429960239" watchObservedRunningTime="2026-03-13 08:01:40.775448332 +0000 UTC m=+1360.446227324" Mar 13 08:01:40 crc kubenswrapper[4876]: I0313 08:01:40.970827 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8dxbx"] Mar 13 08:01:40 crc kubenswrapper[4876]: I0313 08:01:40.980834 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.084024 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc8l4\" (UniqueName: \"kubernetes.io/projected/0888148b-128d-45c4-b6eb-20ce8d458e8c-kube-api-access-wc8l4\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.084206 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.084270 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.084398 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-config\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.084428 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.084518 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.144540 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8dxbx"] Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.187957 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.188614 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc8l4\" (UniqueName: \"kubernetes.io/projected/0888148b-128d-45c4-b6eb-20ce8d458e8c-kube-api-access-wc8l4\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.188745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.188797 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.188878 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-config\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.188908 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.190520 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.193403 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.194508 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.195122 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.195738 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-config\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.220020 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc8l4\" (UniqueName: \"kubernetes.io/projected/0888148b-128d-45c4-b6eb-20ce8d458e8c-kube-api-access-wc8l4\") pod \"dnsmasq-dns-89c5cd4d5-8dxbx\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.311951 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.452646 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:41 crc kubenswrapper[4876]: I0313 08:01:41.757819 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerStarted","Data":"f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa"} Mar 13 08:01:42 crc kubenswrapper[4876]: I0313 08:01:42.038886 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8dxbx"] Mar 13 08:01:42 crc kubenswrapper[4876]: W0313 08:01:42.052280 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0888148b_128d_45c4_b6eb_20ce8d458e8c.slice/crio-97c755411b53be6df78bb8b5ae422227f1ae28988f997de61f91409cf84cbedd WatchSource:0}: Error finding container 97c755411b53be6df78bb8b5ae422227f1ae28988f997de61f91409cf84cbedd: Status 404 returned error can't find the container with id 97c755411b53be6df78bb8b5ae422227f1ae28988f997de61f91409cf84cbedd Mar 13 08:01:42 crc kubenswrapper[4876]: I0313 08:01:42.771424 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerStarted","Data":"4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea"} Mar 13 08:01:42 crc kubenswrapper[4876]: I0313 08:01:42.774034 4876 generic.go:334] "Generic (PLEG): container finished" podID="0888148b-128d-45c4-b6eb-20ce8d458e8c" containerID="5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6" exitCode=0 Mar 13 08:01:42 crc kubenswrapper[4876]: I0313 08:01:42.774139 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" event={"ID":"0888148b-128d-45c4-b6eb-20ce8d458e8c","Type":"ContainerDied","Data":"5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6"} Mar 13 08:01:42 crc kubenswrapper[4876]: I0313 08:01:42.774209 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" event={"ID":"0888148b-128d-45c4-b6eb-20ce8d458e8c","Type":"ContainerStarted","Data":"97c755411b53be6df78bb8b5ae422227f1ae28988f997de61f91409cf84cbedd"} Mar 13 08:01:43 crc kubenswrapper[4876]: I0313 08:01:43.682047 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:43 crc kubenswrapper[4876]: I0313 08:01:43.789629 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" event={"ID":"0888148b-128d-45c4-b6eb-20ce8d458e8c","Type":"ContainerStarted","Data":"7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9"} Mar 13 08:01:43 crc kubenswrapper[4876]: I0313 08:01:43.789841 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-log" containerID="cri-o://afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847" gracePeriod=30 Mar 13 08:01:43 crc kubenswrapper[4876]: I0313 08:01:43.789890 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-api" containerID="cri-o://0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676" gracePeriod=30 Mar 13 08:01:43 crc kubenswrapper[4876]: I0313 08:01:43.869595 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" podStartSLOduration=3.869566184 podStartE2EDuration="3.869566184s" podCreationTimestamp="2026-03-13 08:01:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:43.866374595 +0000 UTC m=+1363.537153577" watchObservedRunningTime="2026-03-13 08:01:43.869566184 +0000 UTC m=+1363.540345166" Mar 13 08:01:44 crc kubenswrapper[4876]: I0313 08:01:44.150982 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:44 crc kubenswrapper[4876]: I0313 08:01:44.806184 4876 generic.go:334] "Generic (PLEG): container finished" podID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerID="afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847" exitCode=143 Mar 13 08:01:44 crc kubenswrapper[4876]: I0313 08:01:44.806299 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ca6048-fb47-45bb-9bc0-a441d684c798","Type":"ContainerDied","Data":"afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847"} Mar 13 08:01:44 crc kubenswrapper[4876]: I0313 08:01:44.812288 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerStarted","Data":"90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c"} Mar 13 08:01:44 crc kubenswrapper[4876]: I0313 08:01:44.812342 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:44 crc kubenswrapper[4876]: I0313 08:01:44.812379 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 08:01:44 crc kubenswrapper[4876]: I0313 08:01:44.846419 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.4714163239999998 podStartE2EDuration="6.846382604s" podCreationTimestamp="2026-03-13 08:01:38 +0000 UTC" firstStartedPulling="2026-03-13 08:01:39.809127501 +0000 UTC m=+1359.479906483" lastFinishedPulling="2026-03-13 08:01:44.184093781 +0000 UTC m=+1363.854872763" observedRunningTime="2026-03-13 08:01:44.833797631 +0000 UTC m=+1364.504576613" watchObservedRunningTime="2026-03-13 08:01:44.846382604 +0000 UTC m=+1364.517161586" Mar 13 08:01:45 crc kubenswrapper[4876]: I0313 08:01:45.008230 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:46 crc kubenswrapper[4876]: I0313 08:01:46.837274 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="ceilometer-central-agent" containerID="cri-o://bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf" gracePeriod=30 Mar 13 08:01:46 crc kubenswrapper[4876]: I0313 08:01:46.837545 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="sg-core" containerID="cri-o://4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea" gracePeriod=30 Mar 13 08:01:46 crc kubenswrapper[4876]: I0313 08:01:46.837617 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="proxy-httpd" containerID="cri-o://90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c" gracePeriod=30 Mar 13 08:01:46 crc kubenswrapper[4876]: I0313 08:01:46.837653 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="ceilometer-notification-agent" containerID="cri-o://f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa" gracePeriod=30 Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.432085 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.533790 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ca6048-fb47-45bb-9bc0-a441d684c798-logs\") pod \"23ca6048-fb47-45bb-9bc0-a441d684c798\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.533955 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72w8c\" (UniqueName: \"kubernetes.io/projected/23ca6048-fb47-45bb-9bc0-a441d684c798-kube-api-access-72w8c\") pod \"23ca6048-fb47-45bb-9bc0-a441d684c798\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.534053 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-config-data\") pod \"23ca6048-fb47-45bb-9bc0-a441d684c798\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.534114 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-combined-ca-bundle\") pod \"23ca6048-fb47-45bb-9bc0-a441d684c798\" (UID: \"23ca6048-fb47-45bb-9bc0-a441d684c798\") " Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.534955 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23ca6048-fb47-45bb-9bc0-a441d684c798-logs" (OuterVolumeSpecName: "logs") pod "23ca6048-fb47-45bb-9bc0-a441d684c798" (UID: "23ca6048-fb47-45bb-9bc0-a441d684c798"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.540556 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ca6048-fb47-45bb-9bc0-a441d684c798-kube-api-access-72w8c" (OuterVolumeSpecName: "kube-api-access-72w8c") pod "23ca6048-fb47-45bb-9bc0-a441d684c798" (UID: "23ca6048-fb47-45bb-9bc0-a441d684c798"). InnerVolumeSpecName "kube-api-access-72w8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.572911 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23ca6048-fb47-45bb-9bc0-a441d684c798" (UID: "23ca6048-fb47-45bb-9bc0-a441d684c798"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.575415 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-config-data" (OuterVolumeSpecName: "config-data") pod "23ca6048-fb47-45bb-9bc0-a441d684c798" (UID: "23ca6048-fb47-45bb-9bc0-a441d684c798"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.636693 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ca6048-fb47-45bb-9bc0-a441d684c798-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.636741 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72w8c\" (UniqueName: \"kubernetes.io/projected/23ca6048-fb47-45bb-9bc0-a441d684c798-kube-api-access-72w8c\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.636756 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.636765 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ca6048-fb47-45bb-9bc0-a441d684c798-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.851111 4876 generic.go:334] "Generic (PLEG): container finished" podID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerID="0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676" exitCode=0 Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.851196 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ca6048-fb47-45bb-9bc0-a441d684c798","Type":"ContainerDied","Data":"0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676"} Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.851233 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ca6048-fb47-45bb-9bc0-a441d684c798","Type":"ContainerDied","Data":"84192caa1c00be3c5ae6844baf19e16845969b93f2b04a9a4f35b56ce26ae49a"} Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.851297 4876 scope.go:117] "RemoveContainer" containerID="0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.851460 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.859341 4876 generic.go:334] "Generic (PLEG): container finished" podID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerID="90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c" exitCode=0 Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.859405 4876 generic.go:334] "Generic (PLEG): container finished" podID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerID="4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea" exitCode=2 Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.859413 4876 generic.go:334] "Generic (PLEG): container finished" podID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerID="f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa" exitCode=0 Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.859411 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerDied","Data":"90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c"} Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.859471 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerDied","Data":"4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea"} Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.859485 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerDied","Data":"f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa"} Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.912041 4876 scope.go:117] "RemoveContainer" containerID="afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.950222 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.971382 4876 scope.go:117] "RemoveContainer" containerID="0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676" Mar 13 08:01:47 crc kubenswrapper[4876]: E0313 08:01:47.972295 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676\": container with ID starting with 0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676 not found: ID does not exist" containerID="0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.972371 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676"} err="failed to get container status \"0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676\": rpc error: code = NotFound desc = could not find container \"0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676\": container with ID starting with 0fab2de358e7a3acb398af9dee830c2d5dea14e6773a1148b44d363a97538676 not found: ID does not exist" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.972405 4876 scope.go:117] "RemoveContainer" containerID="afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847" Mar 13 08:01:47 crc kubenswrapper[4876]: E0313 08:01:47.973044 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847\": container with ID starting with afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847 not found: ID does not exist" containerID="afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.973082 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847"} err="failed to get container status \"afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847\": rpc error: code = NotFound desc = could not find container \"afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847\": container with ID starting with afb11b162b9e28dabae227b300ac68db36e16fe82dd75b202ca6adcd99335847 not found: ID does not exist" Mar 13 08:01:47 crc kubenswrapper[4876]: I0313 08:01:47.981281 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.000327 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:48 crc kubenswrapper[4876]: E0313 08:01:48.001466 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-api" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.001492 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-api" Mar 13 08:01:48 crc kubenswrapper[4876]: E0313 08:01:48.001547 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-log" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.001554 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-log" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.002018 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-api" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.002054 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" containerName="nova-api-log" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.004183 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.008758 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.011108 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.012332 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.016410 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.154807 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.155348 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-public-tls-certs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.155430 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.155579 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf56t\" (UniqueName: \"kubernetes.io/projected/75f0667f-99a8-4846-8d11-32a6bf94b83b-kube-api-access-bf56t\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.156189 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75f0667f-99a8-4846-8d11-32a6bf94b83b-logs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.156423 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-config-data\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.258839 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-config-data\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.258991 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.259035 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-public-tls-certs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.259078 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.259138 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf56t\" (UniqueName: \"kubernetes.io/projected/75f0667f-99a8-4846-8d11-32a6bf94b83b-kube-api-access-bf56t\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.259228 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75f0667f-99a8-4846-8d11-32a6bf94b83b-logs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.260537 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75f0667f-99a8-4846-8d11-32a6bf94b83b-logs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.265662 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-public-tls-certs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.265956 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.266119 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.267789 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-config-data\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.279571 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf56t\" (UniqueName: \"kubernetes.io/projected/75f0667f-99a8-4846-8d11-32a6bf94b83b-kube-api-access-bf56t\") pod \"nova-api-0\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.350810 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:01:48 crc kubenswrapper[4876]: I0313 08:01:48.886129 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:01:48 crc kubenswrapper[4876]: W0313 08:01:48.887863 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75f0667f_99a8_4846_8d11_32a6bf94b83b.slice/crio-66dd6f711e7db6c66157df3fb42183e86b8d16f06e40391fddbea57b96080204 WatchSource:0}: Error finding container 66dd6f711e7db6c66157df3fb42183e86b8d16f06e40391fddbea57b96080204: Status 404 returned error can't find the container with id 66dd6f711e7db6c66157df3fb42183e86b8d16f06e40391fddbea57b96080204 Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.049282 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ca6048-fb47-45bb-9bc0-a441d684c798" path="/var/lib/kubelet/pods/23ca6048-fb47-45bb-9bc0-a441d684c798/volumes" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.151363 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.173657 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.478906 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.538146 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-scripts\") pod \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.538208 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-run-httpd\") pod \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.538347 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-log-httpd\") pod \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.538429 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-sg-core-conf-yaml\") pod \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.538535 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-ceilometer-tls-certs\") pod \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.538551 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-config-data\") pod \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.538641 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-combined-ca-bundle\") pod \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.538701 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4qvn\" (UniqueName: \"kubernetes.io/projected/f5cdea04-68df-432d-8cd5-9b6f5cba9712-kube-api-access-v4qvn\") pod \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\" (UID: \"f5cdea04-68df-432d-8cd5-9b6f5cba9712\") " Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.545706 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-scripts" (OuterVolumeSpecName: "scripts") pod "f5cdea04-68df-432d-8cd5-9b6f5cba9712" (UID: "f5cdea04-68df-432d-8cd5-9b6f5cba9712"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.545966 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5cdea04-68df-432d-8cd5-9b6f5cba9712-kube-api-access-v4qvn" (OuterVolumeSpecName: "kube-api-access-v4qvn") pod "f5cdea04-68df-432d-8cd5-9b6f5cba9712" (UID: "f5cdea04-68df-432d-8cd5-9b6f5cba9712"). InnerVolumeSpecName "kube-api-access-v4qvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.546867 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f5cdea04-68df-432d-8cd5-9b6f5cba9712" (UID: "f5cdea04-68df-432d-8cd5-9b6f5cba9712"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.548413 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f5cdea04-68df-432d-8cd5-9b6f5cba9712" (UID: "f5cdea04-68df-432d-8cd5-9b6f5cba9712"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.602710 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f5cdea04-68df-432d-8cd5-9b6f5cba9712" (UID: "f5cdea04-68df-432d-8cd5-9b6f5cba9712"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.635118 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f5cdea04-68df-432d-8cd5-9b6f5cba9712" (UID: "f5cdea04-68df-432d-8cd5-9b6f5cba9712"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.641671 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.641707 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.641721 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5cdea04-68df-432d-8cd5-9b6f5cba9712-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.641735 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.641745 4876 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.641755 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4qvn\" (UniqueName: \"kubernetes.io/projected/f5cdea04-68df-432d-8cd5-9b6f5cba9712-kube-api-access-v4qvn\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.673102 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-config-data" (OuterVolumeSpecName: "config-data") pod "f5cdea04-68df-432d-8cd5-9b6f5cba9712" (UID: "f5cdea04-68df-432d-8cd5-9b6f5cba9712"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.676073 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5cdea04-68df-432d-8cd5-9b6f5cba9712" (UID: "f5cdea04-68df-432d-8cd5-9b6f5cba9712"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.743793 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.743827 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5cdea04-68df-432d-8cd5-9b6f5cba9712-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.891612 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75f0667f-99a8-4846-8d11-32a6bf94b83b","Type":"ContainerStarted","Data":"4b791e896a8d771005979cef57e910c008e7ac361f29800c3e151a8fbeca0751"} Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.891678 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75f0667f-99a8-4846-8d11-32a6bf94b83b","Type":"ContainerStarted","Data":"81f6aba628fb9a241fd41b029fbc0358f3eb5a533fbdcebba9a35a9c734df151"} Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.891691 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75f0667f-99a8-4846-8d11-32a6bf94b83b","Type":"ContainerStarted","Data":"66dd6f711e7db6c66157df3fb42183e86b8d16f06e40391fddbea57b96080204"} Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.895161 4876 generic.go:334] "Generic (PLEG): container finished" podID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerID="bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf" exitCode=0 Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.895281 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerDied","Data":"bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf"} Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.895317 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.895357 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5cdea04-68df-432d-8cd5-9b6f5cba9712","Type":"ContainerDied","Data":"aead3067c9cbddb025ebc154af04a221278bcc22b0329c38bb1361135ce01520"} Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.895381 4876 scope.go:117] "RemoveContainer" containerID="90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.925315 4876 scope.go:117] "RemoveContainer" containerID="4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.930695 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.930669632 podStartE2EDuration="2.930669632s" podCreationTimestamp="2026-03-13 08:01:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:49.916818983 +0000 UTC m=+1369.587597965" watchObservedRunningTime="2026-03-13 08:01:49.930669632 +0000 UTC m=+1369.601448614" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.931536 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.954296 4876 scope.go:117] "RemoveContainer" containerID="f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa" Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.956935 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:49 crc kubenswrapper[4876]: I0313 08:01:49.988692 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.012321 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:50 crc kubenswrapper[4876]: E0313 08:01:50.012864 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="sg-core" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.012889 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="sg-core" Mar 13 08:01:50 crc kubenswrapper[4876]: E0313 08:01:50.012913 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="proxy-httpd" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.012922 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="proxy-httpd" Mar 13 08:01:50 crc kubenswrapper[4876]: E0313 08:01:50.012944 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="ceilometer-notification-agent" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.012951 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="ceilometer-notification-agent" Mar 13 08:01:50 crc kubenswrapper[4876]: E0313 08:01:50.012963 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="ceilometer-central-agent" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.012970 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="ceilometer-central-agent" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.013158 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="proxy-httpd" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.013175 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="ceilometer-notification-agent" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.013188 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="sg-core" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.013208 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" containerName="ceilometer-central-agent" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.015190 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.017523 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.019638 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.024005 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.068425 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.084320 4876 scope.go:117] "RemoveContainer" containerID="bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.116658 4876 scope.go:117] "RemoveContainer" containerID="90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c" Mar 13 08:01:50 crc kubenswrapper[4876]: E0313 08:01:50.117264 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c\": container with ID starting with 90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c not found: ID does not exist" containerID="90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.117310 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c"} err="failed to get container status \"90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c\": rpc error: code = NotFound desc = could not find container \"90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c\": container with ID starting with 90c0f4bb2e7dbd1793534bc94365003533c6fd010b3c4b03898e5ee16201236c not found: ID does not exist" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.117335 4876 scope.go:117] "RemoveContainer" containerID="4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea" Mar 13 08:01:50 crc kubenswrapper[4876]: E0313 08:01:50.117561 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea\": container with ID starting with 4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea not found: ID does not exist" containerID="4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.117594 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea"} err="failed to get container status \"4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea\": rpc error: code = NotFound desc = could not find container \"4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea\": container with ID starting with 4f81b97feac7a838092fa6ae72ac65997446367f01fcd7fb58caecdece7257ea not found: ID does not exist" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.117606 4876 scope.go:117] "RemoveContainer" containerID="f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa" Mar 13 08:01:50 crc kubenswrapper[4876]: E0313 08:01:50.118437 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa\": container with ID starting with f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa not found: ID does not exist" containerID="f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.118464 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa"} err="failed to get container status \"f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa\": rpc error: code = NotFound desc = could not find container \"f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa\": container with ID starting with f256d2bcccd18d5c38dd798f219b34bd996154b0c23bb44d786a358b64ea22fa not found: ID does not exist" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.118482 4876 scope.go:117] "RemoveContainer" containerID="bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf" Mar 13 08:01:50 crc kubenswrapper[4876]: E0313 08:01:50.118689 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf\": container with ID starting with bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf not found: ID does not exist" containerID="bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.118713 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf"} err="failed to get container status \"bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf\": rpc error: code = NotFound desc = could not find container \"bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf\": container with ID starting with bac301dd810ae9de462de3534a7331bebc1a035cf9faf327e4b3400d423d03bf not found: ID does not exist" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.182009 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27e28286-b51c-43cf-acbb-d612cae68530-log-httpd\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.182063 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.182112 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27e28286-b51c-43cf-acbb-d612cae68530-run-httpd\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.182134 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.182181 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp5sx\" (UniqueName: \"kubernetes.io/projected/27e28286-b51c-43cf-acbb-d612cae68530-kube-api-access-zp5sx\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.182213 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-scripts\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.182463 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-config-data\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.182593 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.184155 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-4w2ft"] Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.185956 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.189994 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.190300 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.196336 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4w2ft"] Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285255 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp5sx\" (UniqueName: \"kubernetes.io/projected/27e28286-b51c-43cf-acbb-d612cae68530-kube-api-access-zp5sx\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285331 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-scripts\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285390 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-config-data\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285424 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285472 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-config-data\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285496 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-scripts\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285517 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27e28286-b51c-43cf-acbb-d612cae68530-log-httpd\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285535 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285558 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285694 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27e28286-b51c-43cf-acbb-d612cae68530-run-httpd\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285720 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.285754 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7shgg\" (UniqueName: \"kubernetes.io/projected/41ef81ac-d481-4e47-81d1-c4846012caae-kube-api-access-7shgg\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.286915 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27e28286-b51c-43cf-acbb-d612cae68530-run-httpd\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.286954 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27e28286-b51c-43cf-acbb-d612cae68530-log-httpd\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.290833 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-config-data\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.291204 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-scripts\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.291228 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.291851 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.292456 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/27e28286-b51c-43cf-acbb-d612cae68530-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.301987 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp5sx\" (UniqueName: \"kubernetes.io/projected/27e28286-b51c-43cf-acbb-d612cae68530-kube-api-access-zp5sx\") pod \"ceilometer-0\" (UID: \"27e28286-b51c-43cf-acbb-d612cae68530\") " pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.382039 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.388350 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7shgg\" (UniqueName: \"kubernetes.io/projected/41ef81ac-d481-4e47-81d1-c4846012caae-kube-api-access-7shgg\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.388608 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-config-data\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.388666 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-scripts\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.388700 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.393704 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.394796 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-config-data\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.395382 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-scripts\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.410478 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7shgg\" (UniqueName: \"kubernetes.io/projected/41ef81ac-d481-4e47-81d1-c4846012caae-kube-api-access-7shgg\") pod \"nova-cell1-cell-mapping-4w2ft\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.512660 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:50 crc kubenswrapper[4876]: I0313 08:01:50.894621 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.048618 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5cdea04-68df-432d-8cd5-9b6f5cba9712" path="/var/lib/kubelet/pods/f5cdea04-68df-432d-8cd5-9b6f5cba9712/volumes" Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.049694 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4w2ft"] Mar 13 08:01:51 crc kubenswrapper[4876]: W0313 08:01:51.051114 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41ef81ac_d481_4e47_81d1_c4846012caae.slice/crio-3c05f14b1107b50214b90c2f2ae4b8df825718c8248a69f5ad54cb471635f3d9 WatchSource:0}: Error finding container 3c05f14b1107b50214b90c2f2ae4b8df825718c8248a69f5ad54cb471635f3d9: Status 404 returned error can't find the container with id 3c05f14b1107b50214b90c2f2ae4b8df825718c8248a69f5ad54cb471635f3d9 Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.453981 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.541302 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-47hc5"] Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.542078 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" podUID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" containerName="dnsmasq-dns" containerID="cri-o://54d08cc86dafab14b9d50bc909b298cea46af95983be2fded4bf1c1c28665452" gracePeriod=10 Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.929082 4876 generic.go:334] "Generic (PLEG): container finished" podID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" containerID="54d08cc86dafab14b9d50bc909b298cea46af95983be2fded4bf1c1c28665452" exitCode=0 Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.929166 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" event={"ID":"44b18fd5-828d-41ce-a174-b7b116d9e8bb","Type":"ContainerDied","Data":"54d08cc86dafab14b9d50bc909b298cea46af95983be2fded4bf1c1c28665452"} Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.933140 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27e28286-b51c-43cf-acbb-d612cae68530","Type":"ContainerStarted","Data":"b489142b8572621e0bb897cf27562a92e3ee065d49d7dabfb30e8b3b0aa42927"} Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.933187 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27e28286-b51c-43cf-acbb-d612cae68530","Type":"ContainerStarted","Data":"6fb5bfa01126264fed805cde224f951cff5206829b9ce2184b0b4e8140d04ec7"} Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.937026 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4w2ft" event={"ID":"41ef81ac-d481-4e47-81d1-c4846012caae","Type":"ContainerStarted","Data":"056f5011fdbb1d4f4c1dc934977d34db0e3d9741d5b5853de942e1a8a0b6e872"} Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.937087 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4w2ft" event={"ID":"41ef81ac-d481-4e47-81d1-c4846012caae","Type":"ContainerStarted","Data":"3c05f14b1107b50214b90c2f2ae4b8df825718c8248a69f5ad54cb471635f3d9"} Mar 13 08:01:51 crc kubenswrapper[4876]: I0313 08:01:51.974096 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-4w2ft" podStartSLOduration=1.974067297 podStartE2EDuration="1.974067297s" podCreationTimestamp="2026-03-13 08:01:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:01:51.959540239 +0000 UTC m=+1371.630319221" watchObservedRunningTime="2026-03-13 08:01:51.974067297 +0000 UTC m=+1371.644846279" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.198286 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.246908 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f77x\" (UniqueName: \"kubernetes.io/projected/44b18fd5-828d-41ce-a174-b7b116d9e8bb-kube-api-access-2f77x\") pod \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.246979 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-config\") pod \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.247125 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-swift-storage-0\") pod \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.247267 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-svc\") pod \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.247289 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-sb\") pod \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.247384 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-nb\") pod \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\" (UID: \"44b18fd5-828d-41ce-a174-b7b116d9e8bb\") " Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.262097 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b18fd5-828d-41ce-a174-b7b116d9e8bb-kube-api-access-2f77x" (OuterVolumeSpecName: "kube-api-access-2f77x") pod "44b18fd5-828d-41ce-a174-b7b116d9e8bb" (UID: "44b18fd5-828d-41ce-a174-b7b116d9e8bb"). InnerVolumeSpecName "kube-api-access-2f77x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.349753 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f77x\" (UniqueName: \"kubernetes.io/projected/44b18fd5-828d-41ce-a174-b7b116d9e8bb-kube-api-access-2f77x\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.353378 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-config" (OuterVolumeSpecName: "config") pod "44b18fd5-828d-41ce-a174-b7b116d9e8bb" (UID: "44b18fd5-828d-41ce-a174-b7b116d9e8bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.363480 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "44b18fd5-828d-41ce-a174-b7b116d9e8bb" (UID: "44b18fd5-828d-41ce-a174-b7b116d9e8bb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.371849 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "44b18fd5-828d-41ce-a174-b7b116d9e8bb" (UID: "44b18fd5-828d-41ce-a174-b7b116d9e8bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.380695 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "44b18fd5-828d-41ce-a174-b7b116d9e8bb" (UID: "44b18fd5-828d-41ce-a174-b7b116d9e8bb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.382342 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "44b18fd5-828d-41ce-a174-b7b116d9e8bb" (UID: "44b18fd5-828d-41ce-a174-b7b116d9e8bb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.451691 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.451737 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-config\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.451755 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.451768 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.451780 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b18fd5-828d-41ce-a174-b7b116d9e8bb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.950852 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" event={"ID":"44b18fd5-828d-41ce-a174-b7b116d9e8bb","Type":"ContainerDied","Data":"fad686ba62c961f45d857292ad8e4e3be9b37e75f988126bed23af4d3816eeb8"} Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.953489 4876 scope.go:117] "RemoveContainer" containerID="54d08cc86dafab14b9d50bc909b298cea46af95983be2fded4bf1c1c28665452" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.952549 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:01:52 crc kubenswrapper[4876]: I0313 08:01:52.955544 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27e28286-b51c-43cf-acbb-d612cae68530","Type":"ContainerStarted","Data":"57a72cb856a9642025886f37ee98e74e6d29dd1d88c787e680d730d17b684136"} Mar 13 08:01:53 crc kubenswrapper[4876]: I0313 08:01:53.025317 4876 scope.go:117] "RemoveContainer" containerID="38752a93dd3eeba29aae494d45d90ad83eed588344764fd59b8a0700883cf4c6" Mar 13 08:01:53 crc kubenswrapper[4876]: I0313 08:01:53.969625 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27e28286-b51c-43cf-acbb-d612cae68530","Type":"ContainerStarted","Data":"0ca0334469066491038914ca943a2dadc031fe62bc74c65166fcb1e45f254be3"} Mar 13 08:01:54 crc kubenswrapper[4876]: I0313 08:01:54.607667 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:01:54 crc kubenswrapper[4876]: I0313 08:01:54.608130 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:01:54 crc kubenswrapper[4876]: I0313 08:01:54.985348 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27e28286-b51c-43cf-acbb-d612cae68530","Type":"ContainerStarted","Data":"310c186a1890ab52c9ae1773863da53ffbbc7bb22d951bb229cd3d919a17a5ef"} Mar 13 08:01:54 crc kubenswrapper[4876]: I0313 08:01:54.986750 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 08:01:55 crc kubenswrapper[4876]: I0313 08:01:55.021884 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.304836508 podStartE2EDuration="6.021847303s" podCreationTimestamp="2026-03-13 08:01:49 +0000 UTC" firstStartedPulling="2026-03-13 08:01:50.920512306 +0000 UTC m=+1370.591291288" lastFinishedPulling="2026-03-13 08:01:54.637523101 +0000 UTC m=+1374.308302083" observedRunningTime="2026-03-13 08:01:55.010939038 +0000 UTC m=+1374.681718040" watchObservedRunningTime="2026-03-13 08:01:55.021847303 +0000 UTC m=+1374.692626285" Mar 13 08:01:58 crc kubenswrapper[4876]: I0313 08:01:58.020841 4876 generic.go:334] "Generic (PLEG): container finished" podID="41ef81ac-d481-4e47-81d1-c4846012caae" containerID="056f5011fdbb1d4f4c1dc934977d34db0e3d9741d5b5853de942e1a8a0b6e872" exitCode=0 Mar 13 08:01:58 crc kubenswrapper[4876]: I0313 08:01:58.020935 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4w2ft" event={"ID":"41ef81ac-d481-4e47-81d1-c4846012caae","Type":"ContainerDied","Data":"056f5011fdbb1d4f4c1dc934977d34db0e3d9741d5b5853de942e1a8a0b6e872"} Mar 13 08:01:58 crc kubenswrapper[4876]: I0313 08:01:58.351635 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 08:01:58 crc kubenswrapper[4876]: I0313 08:01:58.352032 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.370394 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.370507 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.502014 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.574196 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7shgg\" (UniqueName: \"kubernetes.io/projected/41ef81ac-d481-4e47-81d1-c4846012caae-kube-api-access-7shgg\") pod \"41ef81ac-d481-4e47-81d1-c4846012caae\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.574414 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-config-data\") pod \"41ef81ac-d481-4e47-81d1-c4846012caae\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.574445 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-combined-ca-bundle\") pod \"41ef81ac-d481-4e47-81d1-c4846012caae\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.574716 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-scripts\") pod \"41ef81ac-d481-4e47-81d1-c4846012caae\" (UID: \"41ef81ac-d481-4e47-81d1-c4846012caae\") " Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.592552 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41ef81ac-d481-4e47-81d1-c4846012caae-kube-api-access-7shgg" (OuterVolumeSpecName: "kube-api-access-7shgg") pod "41ef81ac-d481-4e47-81d1-c4846012caae" (UID: "41ef81ac-d481-4e47-81d1-c4846012caae"). InnerVolumeSpecName "kube-api-access-7shgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.593771 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-scripts" (OuterVolumeSpecName: "scripts") pod "41ef81ac-d481-4e47-81d1-c4846012caae" (UID: "41ef81ac-d481-4e47-81d1-c4846012caae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.611173 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41ef81ac-d481-4e47-81d1-c4846012caae" (UID: "41ef81ac-d481-4e47-81d1-c4846012caae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.629503 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-config-data" (OuterVolumeSpecName: "config-data") pod "41ef81ac-d481-4e47-81d1-c4846012caae" (UID: "41ef81ac-d481-4e47-81d1-c4846012caae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.677438 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.677497 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.677515 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41ef81ac-d481-4e47-81d1-c4846012caae-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 08:01:59 crc kubenswrapper[4876]: I0313 08:01:59.677529 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7shgg\" (UniqueName: \"kubernetes.io/projected/41ef81ac-d481-4e47-81d1-c4846012caae-kube-api-access-7shgg\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.045933 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4w2ft" event={"ID":"41ef81ac-d481-4e47-81d1-c4846012caae","Type":"ContainerDied","Data":"3c05f14b1107b50214b90c2f2ae4b8df825718c8248a69f5ad54cb471635f3d9"} Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.046608 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c05f14b1107b50214b90c2f2ae4b8df825718c8248a69f5ad54cb471635f3d9" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.046030 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4w2ft" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.160638 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556482-85shb"] Mar 13 08:02:00 crc kubenswrapper[4876]: E0313 08:02:00.161336 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" containerName="init" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.161363 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" containerName="init" Mar 13 08:02:00 crc kubenswrapper[4876]: E0313 08:02:00.161376 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" containerName="dnsmasq-dns" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.161385 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" containerName="dnsmasq-dns" Mar 13 08:02:00 crc kubenswrapper[4876]: E0313 08:02:00.161423 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ef81ac-d481-4e47-81d1-c4846012caae" containerName="nova-manage" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.161432 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ef81ac-d481-4e47-81d1-c4846012caae" containerName="nova-manage" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.161659 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" containerName="dnsmasq-dns" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.161681 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ef81ac-d481-4e47-81d1-c4846012caae" containerName="nova-manage" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.162673 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556482-85shb" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.167314 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.167996 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.168323 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.187341 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556482-85shb"] Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.259735 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.260016 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-log" containerID="cri-o://81f6aba628fb9a241fd41b029fbc0358f3eb5a533fbdcebba9a35a9c734df151" gracePeriod=30 Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.260194 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-api" containerID="cri-o://4b791e896a8d771005979cef57e910c008e7ac361f29800c3e151a8fbeca0751" gracePeriod=30 Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.287400 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.287757 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1da86b72-7254-41cf-a337-c5f309865f27" containerName="nova-scheduler-scheduler" containerID="cri-o://db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541" gracePeriod=30 Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.292032 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv2td\" (UniqueName: \"kubernetes.io/projected/d26e6dde-879c-4a43-98a4-cf908aa696f7-kube-api-access-hv2td\") pod \"auto-csr-approver-29556482-85shb\" (UID: \"d26e6dde-879c-4a43-98a4-cf908aa696f7\") " pod="openshift-infra/auto-csr-approver-29556482-85shb" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.318042 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.319175 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-log" containerID="cri-o://fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c" gracePeriod=30 Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.319838 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-metadata" containerID="cri-o://b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a" gracePeriod=30 Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.393820 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv2td\" (UniqueName: \"kubernetes.io/projected/d26e6dde-879c-4a43-98a4-cf908aa696f7-kube-api-access-hv2td\") pod \"auto-csr-approver-29556482-85shb\" (UID: \"d26e6dde-879c-4a43-98a4-cf908aa696f7\") " pod="openshift-infra/auto-csr-approver-29556482-85shb" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.421649 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv2td\" (UniqueName: \"kubernetes.io/projected/d26e6dde-879c-4a43-98a4-cf908aa696f7-kube-api-access-hv2td\") pod \"auto-csr-approver-29556482-85shb\" (UID: \"d26e6dde-879c-4a43-98a4-cf908aa696f7\") " pod="openshift-infra/auto-csr-approver-29556482-85shb" Mar 13 08:02:00 crc kubenswrapper[4876]: I0313 08:02:00.491741 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556482-85shb" Mar 13 08:02:00 crc kubenswrapper[4876]: E0313 08:02:00.861828 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 08:02:00 crc kubenswrapper[4876]: E0313 08:02:00.864583 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 08:02:00 crc kubenswrapper[4876]: E0313 08:02:00.866688 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 08:02:00 crc kubenswrapper[4876]: E0313 08:02:00.866741 4876 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1da86b72-7254-41cf-a337-c5f309865f27" containerName="nova-scheduler-scheduler" Mar 13 08:02:01 crc kubenswrapper[4876]: I0313 08:02:01.011446 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556482-85shb"] Mar 13 08:02:01 crc kubenswrapper[4876]: I0313 08:02:01.080713 4876 generic.go:334] "Generic (PLEG): container finished" podID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerID="81f6aba628fb9a241fd41b029fbc0358f3eb5a533fbdcebba9a35a9c734df151" exitCode=143 Mar 13 08:02:01 crc kubenswrapper[4876]: I0313 08:02:01.080862 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75f0667f-99a8-4846-8d11-32a6bf94b83b","Type":"ContainerDied","Data":"81f6aba628fb9a241fd41b029fbc0358f3eb5a533fbdcebba9a35a9c734df151"} Mar 13 08:02:01 crc kubenswrapper[4876]: I0313 08:02:01.083854 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556482-85shb" event={"ID":"d26e6dde-879c-4a43-98a4-cf908aa696f7","Type":"ContainerStarted","Data":"7a1c6259a0f8782f6c29af6b557cf94b32ce0240b9928e396409f945cc1dd718"} Mar 13 08:02:01 crc kubenswrapper[4876]: I0313 08:02:01.093396 4876 generic.go:334] "Generic (PLEG): container finished" podID="020dff52-bda7-470d-8e33-7faf45341286" containerID="fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c" exitCode=143 Mar 13 08:02:01 crc kubenswrapper[4876]: I0313 08:02:01.096736 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"020dff52-bda7-470d-8e33-7faf45341286","Type":"ContainerDied","Data":"fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c"} Mar 13 08:02:03 crc kubenswrapper[4876]: I0313 08:02:03.123889 4876 generic.go:334] "Generic (PLEG): container finished" podID="d26e6dde-879c-4a43-98a4-cf908aa696f7" containerID="ab93520b745adca95f5146d9af5af7c14cf6266c8ba8465da18f8188ec64d2b0" exitCode=0 Mar 13 08:02:03 crc kubenswrapper[4876]: I0313 08:02:03.123973 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556482-85shb" event={"ID":"d26e6dde-879c-4a43-98a4-cf908aa696f7","Type":"ContainerDied","Data":"ab93520b745adca95f5146d9af5af7c14cf6266c8ba8465da18f8188ec64d2b0"} Mar 13 08:02:03 crc kubenswrapper[4876]: I0313 08:02:03.469319 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:50540->10.217.0.199:8775: read: connection reset by peer" Mar 13 08:02:03 crc kubenswrapper[4876]: I0313 08:02:03.469741 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:50544->10.217.0.199:8775: read: connection reset by peer" Mar 13 08:02:03 crc kubenswrapper[4876]: I0313 08:02:03.944170 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.078444 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-nova-metadata-tls-certs\") pod \"020dff52-bda7-470d-8e33-7faf45341286\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.078496 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-config-data\") pod \"020dff52-bda7-470d-8e33-7faf45341286\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.078541 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/020dff52-bda7-470d-8e33-7faf45341286-logs\") pod \"020dff52-bda7-470d-8e33-7faf45341286\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.078578 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-combined-ca-bundle\") pod \"020dff52-bda7-470d-8e33-7faf45341286\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.078827 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7qfz\" (UniqueName: \"kubernetes.io/projected/020dff52-bda7-470d-8e33-7faf45341286-kube-api-access-n7qfz\") pod \"020dff52-bda7-470d-8e33-7faf45341286\" (UID: \"020dff52-bda7-470d-8e33-7faf45341286\") " Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.080416 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/020dff52-bda7-470d-8e33-7faf45341286-logs" (OuterVolumeSpecName: "logs") pod "020dff52-bda7-470d-8e33-7faf45341286" (UID: "020dff52-bda7-470d-8e33-7faf45341286"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.086307 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/020dff52-bda7-470d-8e33-7faf45341286-kube-api-access-n7qfz" (OuterVolumeSpecName: "kube-api-access-n7qfz") pod "020dff52-bda7-470d-8e33-7faf45341286" (UID: "020dff52-bda7-470d-8e33-7faf45341286"). InnerVolumeSpecName "kube-api-access-n7qfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.126571 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "020dff52-bda7-470d-8e33-7faf45341286" (UID: "020dff52-bda7-470d-8e33-7faf45341286"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.126610 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-config-data" (OuterVolumeSpecName: "config-data") pod "020dff52-bda7-470d-8e33-7faf45341286" (UID: "020dff52-bda7-470d-8e33-7faf45341286"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.145518 4876 generic.go:334] "Generic (PLEG): container finished" podID="020dff52-bda7-470d-8e33-7faf45341286" containerID="b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a" exitCode=0 Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.145598 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"020dff52-bda7-470d-8e33-7faf45341286","Type":"ContainerDied","Data":"b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a"} Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.145651 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"020dff52-bda7-470d-8e33-7faf45341286","Type":"ContainerDied","Data":"8b0d8fe6bf7aa4c2451af6c7de00a3a1afeec4d2e95b19ec2577a4f081f2cb19"} Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.145681 4876 scope.go:117] "RemoveContainer" containerID="b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.146031 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.182798 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7qfz\" (UniqueName: \"kubernetes.io/projected/020dff52-bda7-470d-8e33-7faf45341286-kube-api-access-n7qfz\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.182836 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.182849 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/020dff52-bda7-470d-8e33-7faf45341286-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.182861 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.192032 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "020dff52-bda7-470d-8e33-7faf45341286" (UID: "020dff52-bda7-470d-8e33-7faf45341286"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.226796 4876 scope.go:117] "RemoveContainer" containerID="fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.251909 4876 scope.go:117] "RemoveContainer" containerID="b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a" Mar 13 08:02:04 crc kubenswrapper[4876]: E0313 08:02:04.252939 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a\": container with ID starting with b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a not found: ID does not exist" containerID="b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.252976 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a"} err="failed to get container status \"b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a\": rpc error: code = NotFound desc = could not find container \"b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a\": container with ID starting with b927cddb1bcc127a5479b43e756fd55574d864b71d5c39625dad2cdd15c8316a not found: ID does not exist" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.253038 4876 scope.go:117] "RemoveContainer" containerID="fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c" Mar 13 08:02:04 crc kubenswrapper[4876]: E0313 08:02:04.254463 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c\": container with ID starting with fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c not found: ID does not exist" containerID="fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.254506 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c"} err="failed to get container status \"fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c\": rpc error: code = NotFound desc = could not find container \"fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c\": container with ID starting with fe3f6c763a4d1b44c5d4f5cf1b7aa945fa9135b30af1a103a792ed62206e094c not found: ID does not exist" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.285496 4876 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/020dff52-bda7-470d-8e33-7faf45341286-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.443553 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556482-85shb" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.488192 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.498749 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.525825 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:02:04 crc kubenswrapper[4876]: E0313 08:02:04.526331 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-log" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.526353 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-log" Mar 13 08:02:04 crc kubenswrapper[4876]: E0313 08:02:04.526387 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d26e6dde-879c-4a43-98a4-cf908aa696f7" containerName="oc" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.526394 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d26e6dde-879c-4a43-98a4-cf908aa696f7" containerName="oc" Mar 13 08:02:04 crc kubenswrapper[4876]: E0313 08:02:04.526400 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-metadata" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.526407 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-metadata" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.526613 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-log" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.526632 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d26e6dde-879c-4a43-98a4-cf908aa696f7" containerName="oc" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.526645 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="020dff52-bda7-470d-8e33-7faf45341286" containerName="nova-metadata-metadata" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.527738 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.535131 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.535461 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.542424 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.597769 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv2td\" (UniqueName: \"kubernetes.io/projected/d26e6dde-879c-4a43-98a4-cf908aa696f7-kube-api-access-hv2td\") pod \"d26e6dde-879c-4a43-98a4-cf908aa696f7\" (UID: \"d26e6dde-879c-4a43-98a4-cf908aa696f7\") " Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.602063 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d26e6dde-879c-4a43-98a4-cf908aa696f7-kube-api-access-hv2td" (OuterVolumeSpecName: "kube-api-access-hv2td") pod "d26e6dde-879c-4a43-98a4-cf908aa696f7" (UID: "d26e6dde-879c-4a43-98a4-cf908aa696f7"). InnerVolumeSpecName "kube-api-access-hv2td". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.700842 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33193662-0ae1-4e6d-bcb8-d093ee42b647-logs\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.700948 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.701000 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-config-data\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.701092 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.701135 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bl7k\" (UniqueName: \"kubernetes.io/projected/33193662-0ae1-4e6d-bcb8-d093ee42b647-kube-api-access-6bl7k\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.701303 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv2td\" (UniqueName: \"kubernetes.io/projected/d26e6dde-879c-4a43-98a4-cf908aa696f7-kube-api-access-hv2td\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.802908 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-config-data\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.802985 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.803027 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bl7k\" (UniqueName: \"kubernetes.io/projected/33193662-0ae1-4e6d-bcb8-d093ee42b647-kube-api-access-6bl7k\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.803161 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33193662-0ae1-4e6d-bcb8-d093ee42b647-logs\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.803201 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.804097 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33193662-0ae1-4e6d-bcb8-d093ee42b647-logs\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.807051 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-config-data\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.807617 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.808228 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33193662-0ae1-4e6d-bcb8-d093ee42b647-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.827109 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bl7k\" (UniqueName: \"kubernetes.io/projected/33193662-0ae1-4e6d-bcb8-d093ee42b647-kube-api-access-6bl7k\") pod \"nova-metadata-0\" (UID: \"33193662-0ae1-4e6d-bcb8-d093ee42b647\") " pod="openstack/nova-metadata-0" Mar 13 08:02:04 crc kubenswrapper[4876]: I0313 08:02:04.963716 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.048980 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="020dff52-bda7-470d-8e33-7faf45341286" path="/var/lib/kubelet/pods/020dff52-bda7-470d-8e33-7faf45341286/volumes" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.167955 4876 generic.go:334] "Generic (PLEG): container finished" podID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerID="4b791e896a8d771005979cef57e910c008e7ac361f29800c3e151a8fbeca0751" exitCode=0 Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.168031 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75f0667f-99a8-4846-8d11-32a6bf94b83b","Type":"ContainerDied","Data":"4b791e896a8d771005979cef57e910c008e7ac361f29800c3e151a8fbeca0751"} Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.168063 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75f0667f-99a8-4846-8d11-32a6bf94b83b","Type":"ContainerDied","Data":"66dd6f711e7db6c66157df3fb42183e86b8d16f06e40391fddbea57b96080204"} Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.168075 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66dd6f711e7db6c66157df3fb42183e86b8d16f06e40391fddbea57b96080204" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.169707 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556482-85shb" event={"ID":"d26e6dde-879c-4a43-98a4-cf908aa696f7","Type":"ContainerDied","Data":"7a1c6259a0f8782f6c29af6b557cf94b32ce0240b9928e396409f945cc1dd718"} Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.169724 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a1c6259a0f8782f6c29af6b557cf94b32ce0240b9928e396409f945cc1dd718" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.169788 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556482-85shb" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.192049 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.321410 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf56t\" (UniqueName: \"kubernetes.io/projected/75f0667f-99a8-4846-8d11-32a6bf94b83b-kube-api-access-bf56t\") pod \"75f0667f-99a8-4846-8d11-32a6bf94b83b\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.321832 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-config-data\") pod \"75f0667f-99a8-4846-8d11-32a6bf94b83b\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.321867 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75f0667f-99a8-4846-8d11-32a6bf94b83b-logs\") pod \"75f0667f-99a8-4846-8d11-32a6bf94b83b\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.321939 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-internal-tls-certs\") pod \"75f0667f-99a8-4846-8d11-32a6bf94b83b\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.321969 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-combined-ca-bundle\") pod \"75f0667f-99a8-4846-8d11-32a6bf94b83b\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.322026 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-public-tls-certs\") pod \"75f0667f-99a8-4846-8d11-32a6bf94b83b\" (UID: \"75f0667f-99a8-4846-8d11-32a6bf94b83b\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.323898 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75f0667f-99a8-4846-8d11-32a6bf94b83b-logs" (OuterVolumeSpecName: "logs") pod "75f0667f-99a8-4846-8d11-32a6bf94b83b" (UID: "75f0667f-99a8-4846-8d11-32a6bf94b83b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.328614 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f0667f-99a8-4846-8d11-32a6bf94b83b-kube-api-access-bf56t" (OuterVolumeSpecName: "kube-api-access-bf56t") pod "75f0667f-99a8-4846-8d11-32a6bf94b83b" (UID: "75f0667f-99a8-4846-8d11-32a6bf94b83b"). InnerVolumeSpecName "kube-api-access-bf56t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.350321 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-config-data" (OuterVolumeSpecName: "config-data") pod "75f0667f-99a8-4846-8d11-32a6bf94b83b" (UID: "75f0667f-99a8-4846-8d11-32a6bf94b83b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.354553 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75f0667f-99a8-4846-8d11-32a6bf94b83b" (UID: "75f0667f-99a8-4846-8d11-32a6bf94b83b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.384489 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "75f0667f-99a8-4846-8d11-32a6bf94b83b" (UID: "75f0667f-99a8-4846-8d11-32a6bf94b83b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.388401 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "75f0667f-99a8-4846-8d11-32a6bf94b83b" (UID: "75f0667f-99a8-4846-8d11-32a6bf94b83b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.424423 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf56t\" (UniqueName: \"kubernetes.io/projected/75f0667f-99a8-4846-8d11-32a6bf94b83b-kube-api-access-bf56t\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.424459 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.424472 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75f0667f-99a8-4846-8d11-32a6bf94b83b-logs\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.424481 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.424489 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.424497 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75f0667f-99a8-4846-8d11-32a6bf94b83b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:05 crc kubenswrapper[4876]: W0313 08:02:05.431583 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33193662_0ae1_4e6d_bcb8_d093ee42b647.slice/crio-13014836e06b39d0f7bb7db19eb489aeece5b20cf0a44d24d1d547cee62e7d5e WatchSource:0}: Error finding container 13014836e06b39d0f7bb7db19eb489aeece5b20cf0a44d24d1d547cee62e7d5e: Status 404 returned error can't find the container with id 13014836e06b39d0f7bb7db19eb489aeece5b20cf0a44d24d1d547cee62e7d5e Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.434515 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.529344 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556476-9x566"] Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.538474 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556476-9x566"] Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.744223 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.832475 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-config-data\") pod \"1da86b72-7254-41cf-a337-c5f309865f27\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.832604 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqf8z\" (UniqueName: \"kubernetes.io/projected/1da86b72-7254-41cf-a337-c5f309865f27-kube-api-access-hqf8z\") pod \"1da86b72-7254-41cf-a337-c5f309865f27\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.832652 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-combined-ca-bundle\") pod \"1da86b72-7254-41cf-a337-c5f309865f27\" (UID: \"1da86b72-7254-41cf-a337-c5f309865f27\") " Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.838501 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da86b72-7254-41cf-a337-c5f309865f27-kube-api-access-hqf8z" (OuterVolumeSpecName: "kube-api-access-hqf8z") pod "1da86b72-7254-41cf-a337-c5f309865f27" (UID: "1da86b72-7254-41cf-a337-c5f309865f27"). InnerVolumeSpecName "kube-api-access-hqf8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.877506 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1da86b72-7254-41cf-a337-c5f309865f27" (UID: "1da86b72-7254-41cf-a337-c5f309865f27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.881685 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-config-data" (OuterVolumeSpecName: "config-data") pod "1da86b72-7254-41cf-a337-c5f309865f27" (UID: "1da86b72-7254-41cf-a337-c5f309865f27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.936191 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.936246 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1da86b72-7254-41cf-a337-c5f309865f27-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:05 crc kubenswrapper[4876]: I0313 08:02:05.936271 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqf8z\" (UniqueName: \"kubernetes.io/projected/1da86b72-7254-41cf-a337-c5f309865f27-kube-api-access-hqf8z\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.185974 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33193662-0ae1-4e6d-bcb8-d093ee42b647","Type":"ContainerStarted","Data":"304448965c6609a6080f3e592a92064dc1ffd96b3fcd7609813795fb989d3022"} Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.186060 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33193662-0ae1-4e6d-bcb8-d093ee42b647","Type":"ContainerStarted","Data":"ab8e4f7c841fa5db0133be295370257c8be28486dc2d7067e24db61318489945"} Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.186083 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33193662-0ae1-4e6d-bcb8-d093ee42b647","Type":"ContainerStarted","Data":"13014836e06b39d0f7bb7db19eb489aeece5b20cf0a44d24d1d547cee62e7d5e"} Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.191293 4876 generic.go:334] "Generic (PLEG): container finished" podID="1da86b72-7254-41cf-a337-c5f309865f27" containerID="db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541" exitCode=0 Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.191319 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.191430 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.192077 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1da86b72-7254-41cf-a337-c5f309865f27","Type":"ContainerDied","Data":"db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541"} Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.192186 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1da86b72-7254-41cf-a337-c5f309865f27","Type":"ContainerDied","Data":"ba8c0e33882e94f1dc207256a2dc91bf682746f1f5b21f675a509f90ee09d63f"} Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.192220 4876 scope.go:117] "RemoveContainer" containerID="db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.217209 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.217179448 podStartE2EDuration="2.217179448s" podCreationTimestamp="2026-03-13 08:02:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:02:06.216409217 +0000 UTC m=+1385.887188219" watchObservedRunningTime="2026-03-13 08:02:06.217179448 +0000 UTC m=+1385.887958430" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.224769 4876 scope.go:117] "RemoveContainer" containerID="db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541" Mar 13 08:02:06 crc kubenswrapper[4876]: E0313 08:02:06.225468 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541\": container with ID starting with db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541 not found: ID does not exist" containerID="db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.225514 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541"} err="failed to get container status \"db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541\": rpc error: code = NotFound desc = could not find container \"db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541\": container with ID starting with db2e3b335b1746c9a434ecb4d1bbe94c6cd8d2510a644b3ccd85b19095abe541 not found: ID does not exist" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.279925 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.294288 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.307200 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.320361 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.330315 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:02:06 crc kubenswrapper[4876]: E0313 08:02:06.331135 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-log" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.331161 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-log" Mar 13 08:02:06 crc kubenswrapper[4876]: E0313 08:02:06.331289 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-api" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.331301 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-api" Mar 13 08:02:06 crc kubenswrapper[4876]: E0313 08:02:06.331342 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da86b72-7254-41cf-a337-c5f309865f27" containerName="nova-scheduler-scheduler" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.331352 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da86b72-7254-41cf-a337-c5f309865f27" containerName="nova-scheduler-scheduler" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.331636 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-log" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.331680 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da86b72-7254-41cf-a337-c5f309865f27" containerName="nova-scheduler-scheduler" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.331695 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" containerName="nova-api-api" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.333069 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.336857 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.344378 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.368264 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.370806 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.373902 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.375942 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.380467 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.385488 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.447691 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg9pk\" (UniqueName: \"kubernetes.io/projected/d51bcb66-a417-4369-94df-a59aedd82d1c-kube-api-access-fg9pk\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.447782 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfd0230f-78c9-4fa3-a450-889afd1fb40a-logs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.447830 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.448120 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.448283 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d51bcb66-a417-4369-94df-a59aedd82d1c-config-data\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.448349 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5fcl\" (UniqueName: \"kubernetes.io/projected/cfd0230f-78c9-4fa3-a450-889afd1fb40a-kube-api-access-w5fcl\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.448377 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-public-tls-certs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.448409 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51bcb66-a417-4369-94df-a59aedd82d1c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.448445 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-config-data\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550219 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51bcb66-a417-4369-94df-a59aedd82d1c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550652 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-config-data\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550712 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg9pk\" (UniqueName: \"kubernetes.io/projected/d51bcb66-a417-4369-94df-a59aedd82d1c-kube-api-access-fg9pk\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfd0230f-78c9-4fa3-a450-889afd1fb40a-logs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550771 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550825 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550848 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d51bcb66-a417-4369-94df-a59aedd82d1c-config-data\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550883 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5fcl\" (UniqueName: \"kubernetes.io/projected/cfd0230f-78c9-4fa3-a450-889afd1fb40a-kube-api-access-w5fcl\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.550900 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-public-tls-certs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.551963 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfd0230f-78c9-4fa3-a450-889afd1fb40a-logs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.557977 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.558567 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-config-data\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.558591 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d51bcb66-a417-4369-94df-a59aedd82d1c-config-data\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.559885 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51bcb66-a417-4369-94df-a59aedd82d1c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.562099 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.566778 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfd0230f-78c9-4fa3-a450-889afd1fb40a-public-tls-certs\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.571022 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5fcl\" (UniqueName: \"kubernetes.io/projected/cfd0230f-78c9-4fa3-a450-889afd1fb40a-kube-api-access-w5fcl\") pod \"nova-api-0\" (UID: \"cfd0230f-78c9-4fa3-a450-889afd1fb40a\") " pod="openstack/nova-api-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.571549 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg9pk\" (UniqueName: \"kubernetes.io/projected/d51bcb66-a417-4369-94df-a59aedd82d1c-kube-api-access-fg9pk\") pod \"nova-scheduler-0\" (UID: \"d51bcb66-a417-4369-94df-a59aedd82d1c\") " pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.666987 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 08:02:06 crc kubenswrapper[4876]: I0313 08:02:06.690381 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 08:02:07 crc kubenswrapper[4876]: I0313 08:02:07.046502 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1da86b72-7254-41cf-a337-c5f309865f27" path="/var/lib/kubelet/pods/1da86b72-7254-41cf-a337-c5f309865f27/volumes" Mar 13 08:02:07 crc kubenswrapper[4876]: I0313 08:02:07.047431 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75f0667f-99a8-4846-8d11-32a6bf94b83b" path="/var/lib/kubelet/pods/75f0667f-99a8-4846-8d11-32a6bf94b83b/volumes" Mar 13 08:02:07 crc kubenswrapper[4876]: I0313 08:02:07.048057 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9431e26b-bb79-45a0-a0a2-9e11479cc64b" path="/var/lib/kubelet/pods/9431e26b-bb79-45a0-a0a2-9e11479cc64b/volumes" Mar 13 08:02:07 crc kubenswrapper[4876]: I0313 08:02:07.173191 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 08:02:07 crc kubenswrapper[4876]: W0313 08:02:07.176693 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd51bcb66_a417_4369_94df_a59aedd82d1c.slice/crio-c97aef0dbf6a13efa65cbba82872204a559ba6ae1a1013b401486fa712da2fde WatchSource:0}: Error finding container c97aef0dbf6a13efa65cbba82872204a559ba6ae1a1013b401486fa712da2fde: Status 404 returned error can't find the container with id c97aef0dbf6a13efa65cbba82872204a559ba6ae1a1013b401486fa712da2fde Mar 13 08:02:07 crc kubenswrapper[4876]: I0313 08:02:07.212688 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d51bcb66-a417-4369-94df-a59aedd82d1c","Type":"ContainerStarted","Data":"c97aef0dbf6a13efa65cbba82872204a559ba6ae1a1013b401486fa712da2fde"} Mar 13 08:02:07 crc kubenswrapper[4876]: W0313 08:02:07.275738 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfd0230f_78c9_4fa3_a450_889afd1fb40a.slice/crio-f948c5da6d5881cfba08f9ea7f0a7a4b0316cb185449bfb99f3cd319ba847008 WatchSource:0}: Error finding container f948c5da6d5881cfba08f9ea7f0a7a4b0316cb185449bfb99f3cd319ba847008: Status 404 returned error can't find the container with id f948c5da6d5881cfba08f9ea7f0a7a4b0316cb185449bfb99f3cd319ba847008 Mar 13 08:02:07 crc kubenswrapper[4876]: I0313 08:02:07.277937 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 08:02:08 crc kubenswrapper[4876]: I0313 08:02:08.226922 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cfd0230f-78c9-4fa3-a450-889afd1fb40a","Type":"ContainerStarted","Data":"4b2453af980cf5902d1081ac4c9e240b5a6e78f62fe2e3ea072a60eae44caa00"} Mar 13 08:02:08 crc kubenswrapper[4876]: I0313 08:02:08.227973 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cfd0230f-78c9-4fa3-a450-889afd1fb40a","Type":"ContainerStarted","Data":"a81d2aaa066cb83c3e50b41d9131892fd6dc9e7be934a79d74278f8f00f393f7"} Mar 13 08:02:08 crc kubenswrapper[4876]: I0313 08:02:08.227991 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cfd0230f-78c9-4fa3-a450-889afd1fb40a","Type":"ContainerStarted","Data":"f948c5da6d5881cfba08f9ea7f0a7a4b0316cb185449bfb99f3cd319ba847008"} Mar 13 08:02:08 crc kubenswrapper[4876]: I0313 08:02:08.230215 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d51bcb66-a417-4369-94df-a59aedd82d1c","Type":"ContainerStarted","Data":"5fa2562ef7250eb51a98cfe75e970741ba422c27ad0facbd33cbd8e74b048f50"} Mar 13 08:02:08 crc kubenswrapper[4876]: I0313 08:02:08.256848 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.256817968 podStartE2EDuration="2.256817968s" podCreationTimestamp="2026-03-13 08:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:02:08.254305557 +0000 UTC m=+1387.925084539" watchObservedRunningTime="2026-03-13 08:02:08.256817968 +0000 UTC m=+1387.927596950" Mar 13 08:02:08 crc kubenswrapper[4876]: I0313 08:02:08.280388 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.280362158 podStartE2EDuration="2.280362158s" podCreationTimestamp="2026-03-13 08:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:02:08.274557095 +0000 UTC m=+1387.945336077" watchObservedRunningTime="2026-03-13 08:02:08.280362158 +0000 UTC m=+1387.951141140" Mar 13 08:02:09 crc kubenswrapper[4876]: I0313 08:02:09.963885 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 08:02:09 crc kubenswrapper[4876]: I0313 08:02:09.964283 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 08:02:11 crc kubenswrapper[4876]: I0313 08:02:11.668043 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 08:02:14 crc kubenswrapper[4876]: I0313 08:02:14.964795 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 08:02:14 crc kubenswrapper[4876]: I0313 08:02:14.965742 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 08:02:15 crc kubenswrapper[4876]: I0313 08:02:15.979497 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="33193662-0ae1-4e6d-bcb8-d093ee42b647" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 08:02:15 crc kubenswrapper[4876]: I0313 08:02:15.979559 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="33193662-0ae1-4e6d-bcb8-d093ee42b647" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 08:02:16 crc kubenswrapper[4876]: I0313 08:02:16.667966 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 08:02:16 crc kubenswrapper[4876]: I0313 08:02:16.691632 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 08:02:16 crc kubenswrapper[4876]: I0313 08:02:16.691694 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 08:02:16 crc kubenswrapper[4876]: I0313 08:02:16.704991 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 08:02:17 crc kubenswrapper[4876]: I0313 08:02:17.372438 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 08:02:17 crc kubenswrapper[4876]: I0313 08:02:17.704488 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cfd0230f-78c9-4fa3-a450-889afd1fb40a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 08:02:17 crc kubenswrapper[4876]: I0313 08:02:17.704926 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cfd0230f-78c9-4fa3-a450-889afd1fb40a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 08:02:18 crc kubenswrapper[4876]: I0313 08:02:18.678449 4876 scope.go:117] "RemoveContainer" containerID="ae6a3a4dc748772291167dd4abc6cbf51ceb60eb996b6bb924f1925ce25dc20f" Mar 13 08:02:20 crc kubenswrapper[4876]: I0313 08:02:20.391976 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 08:02:23 crc kubenswrapper[4876]: I0313 08:02:23.105447 4876 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod44b18fd5-828d-41ce-a174-b7b116d9e8bb"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod44b18fd5-828d-41ce-a174-b7b116d9e8bb] : Timed out while waiting for systemd to remove kubepods-besteffort-pod44b18fd5_828d_41ce_a174_b7b116d9e8bb.slice" Mar 13 08:02:23 crc kubenswrapper[4876]: E0313 08:02:23.106416 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod44b18fd5-828d-41ce-a174-b7b116d9e8bb] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod44b18fd5-828d-41ce-a174-b7b116d9e8bb] : Timed out while waiting for systemd to remove kubepods-besteffort-pod44b18fd5_828d_41ce_a174_b7b116d9e8bb.slice" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" podUID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" Mar 13 08:02:23 crc kubenswrapper[4876]: I0313 08:02:23.420576 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-47hc5" Mar 13 08:02:23 crc kubenswrapper[4876]: I0313 08:02:23.469922 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-47hc5"] Mar 13 08:02:23 crc kubenswrapper[4876]: I0313 08:02:23.480114 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-47hc5"] Mar 13 08:02:24 crc kubenswrapper[4876]: I0313 08:02:24.608506 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:02:24 crc kubenswrapper[4876]: I0313 08:02:24.608610 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:02:24 crc kubenswrapper[4876]: I0313 08:02:24.970456 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 08:02:24 crc kubenswrapper[4876]: I0313 08:02:24.972649 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 08:02:24 crc kubenswrapper[4876]: I0313 08:02:24.976645 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 08:02:25 crc kubenswrapper[4876]: I0313 08:02:25.047724 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b18fd5-828d-41ce-a174-b7b116d9e8bb" path="/var/lib/kubelet/pods/44b18fd5-828d-41ce-a174-b7b116d9e8bb/volumes" Mar 13 08:02:25 crc kubenswrapper[4876]: I0313 08:02:25.457969 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 08:02:26 crc kubenswrapper[4876]: I0313 08:02:26.699600 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 08:02:26 crc kubenswrapper[4876]: I0313 08:02:26.700255 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 08:02:26 crc kubenswrapper[4876]: I0313 08:02:26.700849 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 08:02:26 crc kubenswrapper[4876]: I0313 08:02:26.713376 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 08:02:27 crc kubenswrapper[4876]: I0313 08:02:27.468926 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 08:02:27 crc kubenswrapper[4876]: I0313 08:02:27.475758 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 08:02:35 crc kubenswrapper[4876]: I0313 08:02:35.293978 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 08:02:36 crc kubenswrapper[4876]: I0313 08:02:36.880485 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 08:02:40 crc kubenswrapper[4876]: I0313 08:02:40.263162 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerName="rabbitmq" containerID="cri-o://7a2e4794cd7b4ce4ab5ad2b4921db1d2e8ca47414faee7c132f8ec81ad3a2c3c" gracePeriod=604796 Mar 13 08:02:42 crc kubenswrapper[4876]: I0313 08:02:42.201993 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerName="rabbitmq" containerID="cri-o://8d33159caad604faa64957e660287e845db43946fa7402d50ae508453ac633a3" gracePeriod=604795 Mar 13 08:02:44 crc kubenswrapper[4876]: I0313 08:02:44.162207 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Mar 13 08:02:44 crc kubenswrapper[4876]: I0313 08:02:44.534192 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:46.693517 4876 generic.go:334] "Generic (PLEG): container finished" podID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerID="7a2e4794cd7b4ce4ab5ad2b4921db1d2e8ca47414faee7c132f8ec81ad3a2c3c" exitCode=0 Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:46.693944 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678","Type":"ContainerDied","Data":"7a2e4794cd7b4ce4ab5ad2b4921db1d2e8ca47414faee7c132f8ec81ad3a2c3c"} Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:46.912620 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.086846 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-tls\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087489 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087575 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j7p4\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-kube-api-access-4j7p4\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087615 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-server-conf\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087656 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-plugins\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087698 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-erlang-cookie\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087729 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-confd\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087852 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-erlang-cookie-secret\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087902 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-config-data\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.087941 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-pod-info\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.090447 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-plugins-conf\") pod \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\" (UID: \"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678\") " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.095628 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.096137 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.096430 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.099169 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.106139 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-kube-api-access-4j7p4" (OuterVolumeSpecName: "kube-api-access-4j7p4") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "kube-api-access-4j7p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.106871 4876 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.106902 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.106918 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j7p4\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-kube-api-access-4j7p4\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.106932 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.106957 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.107181 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.107676 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-pod-info" (OuterVolumeSpecName: "pod-info") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.112432 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.138840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-config-data" (OuterVolumeSpecName: "config-data") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.202314 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-server-conf" (OuterVolumeSpecName: "server-conf") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.211915 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.211951 4876 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.211966 4876 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.211982 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.211993 4876 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.239460 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.256816 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" (UID: "54290a8c-83ce-4d8b-bc4b-a28c2f0e8678"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.314547 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.314590 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.725114 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.725089 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"54290a8c-83ce-4d8b-bc4b-a28c2f0e8678","Type":"ContainerDied","Data":"3773a980a2d5a23e03d1ec13efadf8a908102dd731491440f7d484762ca9d9d0"} Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.725700 4876 scope.go:117] "RemoveContainer" containerID="7a2e4794cd7b4ce4ab5ad2b4921db1d2e8ca47414faee7c132f8ec81ad3a2c3c" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.772641 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.775157 4876 scope.go:117] "RemoveContainer" containerID="7bef2f513a2c87a154de6467ad78491d761e7c372e49596ffdb8f85720c39dea" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.786919 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.803521 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 08:02:47 crc kubenswrapper[4876]: E0313 08:02:47.804213 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerName="setup-container" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.804229 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerName="setup-container" Mar 13 08:02:47 crc kubenswrapper[4876]: E0313 08:02:47.804284 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerName="rabbitmq" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.804294 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerName="rabbitmq" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.804639 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" containerName="rabbitmq" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.806305 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.813962 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.814034 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.814386 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.814532 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.814566 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.814761 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.815013 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2qq6k" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.824873 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.929942 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930014 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930047 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-server-conf\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930188 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930278 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-config-data\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930314 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nkgc\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-kube-api-access-5nkgc\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930365 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79aa8084-23ef-4730-948d-1b0d6ca25251-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930405 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930457 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930487 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:47 crc kubenswrapper[4876]: I0313 08:02:47.930538 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79aa8084-23ef-4730-948d-1b0d6ca25251-pod-info\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.035941 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.036033 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-config-data\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.036357 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nkgc\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-kube-api-access-5nkgc\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.036834 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79aa8084-23ef-4730-948d-1b0d6ca25251-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.036977 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.037060 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.037097 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.037170 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-config-data\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.037335 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79aa8084-23ef-4730-948d-1b0d6ca25251-pod-info\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.037428 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.037474 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.037504 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-server-conf\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.037732 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.038954 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-server-conf\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.040657 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.041009 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.041460 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79aa8084-23ef-4730-948d-1b0d6ca25251-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.064045 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.064841 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79aa8084-23ef-4730-948d-1b0d6ca25251-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.075987 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79aa8084-23ef-4730-948d-1b0d6ca25251-pod-info\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.104153 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.116155 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nkgc\" (UniqueName: \"kubernetes.io/projected/79aa8084-23ef-4730-948d-1b0d6ca25251-kube-api-access-5nkgc\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.182128 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"79aa8084-23ef-4730-948d-1b0d6ca25251\") " pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.480733 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.759901 4876 generic.go:334] "Generic (PLEG): container finished" podID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerID="8d33159caad604faa64957e660287e845db43946fa7402d50ae508453ac633a3" exitCode=0 Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.759989 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cca33d7d-7758-4f14-9d61-1785e84ea2b8","Type":"ContainerDied","Data":"8d33159caad604faa64957e660287e845db43946fa7402d50ae508453ac633a3"} Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.859760 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.956765 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-confd\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.956969 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cca33d7d-7758-4f14-9d61-1785e84ea2b8-pod-info\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.957150 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-plugins\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.957458 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-config-data\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.957638 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzcmk\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-kube-api-access-bzcmk\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.957857 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cca33d7d-7758-4f14-9d61-1785e84ea2b8-erlang-cookie-secret\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.957970 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-erlang-cookie\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.958067 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-server-conf\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.958575 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.958687 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-plugins-conf\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.958840 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-tls\") pod \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\" (UID: \"cca33d7d-7758-4f14-9d61-1785e84ea2b8\") " Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.957669 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.959713 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.959962 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.960015 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.960044 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.967755 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.967805 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cca33d7d-7758-4f14-9d61-1785e84ea2b8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.970360 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.972377 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-kube-api-access-bzcmk" (OuterVolumeSpecName: "kube-api-access-bzcmk") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "kube-api-access-bzcmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:48 crc kubenswrapper[4876]: I0313 08:02:48.977662 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/cca33d7d-7758-4f14-9d61-1785e84ea2b8-pod-info" (OuterVolumeSpecName: "pod-info") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.001908 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-config-data" (OuterVolumeSpecName: "config-data") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.059180 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54290a8c-83ce-4d8b-bc4b-a28c2f0e8678" path="/var/lib/kubelet/pods/54290a8c-83ce-4d8b-bc4b-a28c2f0e8678/volumes" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.062110 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.062133 4876 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.062143 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.062152 4876 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cca33d7d-7758-4f14-9d61-1785e84ea2b8-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.062163 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.062173 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzcmk\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-kube-api-access-bzcmk\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.062183 4876 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cca33d7d-7758-4f14-9d61-1785e84ea2b8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.097740 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.117297 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-server-conf" (OuterVolumeSpecName: "server-conf") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.128280 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.164104 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.164128 4876 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cca33d7d-7758-4f14-9d61-1785e84ea2b8-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.176931 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "cca33d7d-7758-4f14-9d61-1785e84ea2b8" (UID: "cca33d7d-7758-4f14-9d61-1785e84ea2b8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.266349 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cca33d7d-7758-4f14-9d61-1785e84ea2b8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.566452 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-xw8gl"] Mar 13 08:02:49 crc kubenswrapper[4876]: E0313 08:02:49.568078 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerName="setup-container" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.568146 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerName="setup-container" Mar 13 08:02:49 crc kubenswrapper[4876]: E0313 08:02:49.568213 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerName="rabbitmq" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.568322 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerName="rabbitmq" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.568620 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" containerName="rabbitmq" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.569907 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.574764 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.583876 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-xw8gl"] Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.673597 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-config\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.673687 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.673714 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.673753 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.673776 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.673800 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwwqf\" (UniqueName: \"kubernetes.io/projected/bb7a1abc-aec3-4fab-82f6-9d305576db86-kube-api-access-pwwqf\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.673819 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.775851 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.775905 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.775958 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.775984 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.776008 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwwqf\" (UniqueName: \"kubernetes.io/projected/bb7a1abc-aec3-4fab-82f6-9d305576db86-kube-api-access-pwwqf\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.776035 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.776132 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-config\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.777163 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.777163 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.777464 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.777553 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.777593 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-config\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.777864 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.780978 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"79aa8084-23ef-4730-948d-1b0d6ca25251","Type":"ContainerStarted","Data":"42a577e9b87311a3848d4c861d88c035f1896dbeea01d1c8286d37f5bb5e2d9d"} Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.783179 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cca33d7d-7758-4f14-9d61-1785e84ea2b8","Type":"ContainerDied","Data":"d0d93982bd721aa0e0c61c7c3bcd60b435563cc240c52a9857df0cf9d4371b60"} Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.783216 4876 scope.go:117] "RemoveContainer" containerID="8d33159caad604faa64957e660287e845db43946fa7402d50ae508453ac633a3" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.788351 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.794808 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwwqf\" (UniqueName: \"kubernetes.io/projected/bb7a1abc-aec3-4fab-82f6-9d305576db86-kube-api-access-pwwqf\") pod \"dnsmasq-dns-79bd4cc8c9-xw8gl\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:49 crc kubenswrapper[4876]: I0313 08:02:49.891776 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.211395 4876 scope.go:117] "RemoveContainer" containerID="8bbb2cd4e0422ba5625aa003e75bef5ff7479286bc579a6d22675536aa504525" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.222951 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.248786 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.260502 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.263156 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.267372 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.267429 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.267614 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-csqv6" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.267549 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.269335 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.269518 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.271870 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.278509 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.389653 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-xw8gl"] Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.393489 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.393562 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/147c88b9-db25-435e-9575-edb4efa342e7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.393817 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/147c88b9-db25-435e-9575-edb4efa342e7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.393928 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.393973 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zc98\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-kube-api-access-6zc98\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.394215 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.394334 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.394375 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.394434 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.394478 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.394499 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.496605 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.496691 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.496725 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.496789 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.496818 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.496909 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.496951 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/147c88b9-db25-435e-9575-edb4efa342e7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.496996 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/147c88b9-db25-435e-9575-edb4efa342e7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.497040 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.497068 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zc98\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-kube-api-access-6zc98\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.497119 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.497229 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.497595 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.498012 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.499153 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.500318 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.502253 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147c88b9-db25-435e-9575-edb4efa342e7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.601060 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/147c88b9-db25-435e-9575-edb4efa342e7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.602916 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.603155 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/147c88b9-db25-435e-9575-edb4efa342e7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.603645 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.604975 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zc98\" (UniqueName: \"kubernetes.io/projected/147c88b9-db25-435e-9575-edb4efa342e7-kube-api-access-6zc98\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.642853 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"147c88b9-db25-435e-9575-edb4efa342e7\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.804216 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" event={"ID":"bb7a1abc-aec3-4fab-82f6-9d305576db86","Type":"ContainerStarted","Data":"f291fdd40c026e3e67b0fe039587b0e79e43a32c9eb6870268e7815bbe24c51d"} Mar 13 08:02:50 crc kubenswrapper[4876]: I0313 08:02:50.888966 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:02:51 crc kubenswrapper[4876]: I0313 08:02:51.056439 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cca33d7d-7758-4f14-9d61-1785e84ea2b8" path="/var/lib/kubelet/pods/cca33d7d-7758-4f14-9d61-1785e84ea2b8/volumes" Mar 13 08:02:51 crc kubenswrapper[4876]: I0313 08:02:51.397404 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 08:02:51 crc kubenswrapper[4876]: W0313 08:02:51.400953 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod147c88b9_db25_435e_9575_edb4efa342e7.slice/crio-0cece57abdac40791053ff51840f2024906c069d23acaabf0e82a9410f843f8b WatchSource:0}: Error finding container 0cece57abdac40791053ff51840f2024906c069d23acaabf0e82a9410f843f8b: Status 404 returned error can't find the container with id 0cece57abdac40791053ff51840f2024906c069d23acaabf0e82a9410f843f8b Mar 13 08:02:51 crc kubenswrapper[4876]: I0313 08:02:51.818556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"79aa8084-23ef-4730-948d-1b0d6ca25251","Type":"ContainerStarted","Data":"d8f014b4db9da3bcfc70ebac7e22b9b21887bfd1a56d20f8116edfc89e8c3b9d"} Mar 13 08:02:51 crc kubenswrapper[4876]: I0313 08:02:51.821970 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"147c88b9-db25-435e-9575-edb4efa342e7","Type":"ContainerStarted","Data":"0cece57abdac40791053ff51840f2024906c069d23acaabf0e82a9410f843f8b"} Mar 13 08:02:51 crc kubenswrapper[4876]: I0313 08:02:51.823539 4876 generic.go:334] "Generic (PLEG): container finished" podID="bb7a1abc-aec3-4fab-82f6-9d305576db86" containerID="4cec8ecf0ea3abac126fdd5020118cfd79550c900e56214461c4e3705c9b58ce" exitCode=0 Mar 13 08:02:51 crc kubenswrapper[4876]: I0313 08:02:51.823656 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" event={"ID":"bb7a1abc-aec3-4fab-82f6-9d305576db86","Type":"ContainerDied","Data":"4cec8ecf0ea3abac126fdd5020118cfd79550c900e56214461c4e3705c9b58ce"} Mar 13 08:02:52 crc kubenswrapper[4876]: I0313 08:02:52.837186 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" event={"ID":"bb7a1abc-aec3-4fab-82f6-9d305576db86","Type":"ContainerStarted","Data":"50450a1e0da546a7f2ec5efe72a2a4ce3d0d9154e8b9b05e588df635f5db89ac"} Mar 13 08:02:52 crc kubenswrapper[4876]: I0313 08:02:52.871961 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" podStartSLOduration=3.87193558 podStartE2EDuration="3.87193558s" podCreationTimestamp="2026-03-13 08:02:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:02:52.867915007 +0000 UTC m=+1432.538693989" watchObservedRunningTime="2026-03-13 08:02:52.87193558 +0000 UTC m=+1432.542714562" Mar 13 08:02:53 crc kubenswrapper[4876]: I0313 08:02:53.852305 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"147c88b9-db25-435e-9575-edb4efa342e7","Type":"ContainerStarted","Data":"c0a82e136f698070e4b8a7ca79ff71246b26ef7f3c3d2e4e7b12310a06bc08cd"} Mar 13 08:02:53 crc kubenswrapper[4876]: I0313 08:02:53.852614 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:54 crc kubenswrapper[4876]: I0313 08:02:54.608039 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:02:54 crc kubenswrapper[4876]: I0313 08:02:54.608154 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:02:54 crc kubenswrapper[4876]: I0313 08:02:54.608221 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:02:54 crc kubenswrapper[4876]: I0313 08:02:54.609268 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab9fcd397cad6841be9000d6381846b64e56e1d04c705091fe8c6b9884bbef7d"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:02:54 crc kubenswrapper[4876]: I0313 08:02:54.609332 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://ab9fcd397cad6841be9000d6381846b64e56e1d04c705091fe8c6b9884bbef7d" gracePeriod=600 Mar 13 08:02:54 crc kubenswrapper[4876]: I0313 08:02:54.867119 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="ab9fcd397cad6841be9000d6381846b64e56e1d04c705091fe8c6b9884bbef7d" exitCode=0 Mar 13 08:02:54 crc kubenswrapper[4876]: I0313 08:02:54.867202 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"ab9fcd397cad6841be9000d6381846b64e56e1d04c705091fe8c6b9884bbef7d"} Mar 13 08:02:54 crc kubenswrapper[4876]: I0313 08:02:54.867306 4876 scope.go:117] "RemoveContainer" containerID="8da8f1dc197e5903f65bce5f57eb064acd43a39afee0ba7b7e356d5bf1d3a13e" Mar 13 08:02:55 crc kubenswrapper[4876]: I0313 08:02:55.881586 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2"} Mar 13 08:02:59 crc kubenswrapper[4876]: I0313 08:02:59.895369 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:02:59 crc kubenswrapper[4876]: I0313 08:02:59.966177 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8dxbx"] Mar 13 08:02:59 crc kubenswrapper[4876]: I0313 08:02:59.966607 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" podUID="0888148b-128d-45c4-b6eb-20ce8d458e8c" containerName="dnsmasq-dns" containerID="cri-o://7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9" gracePeriod=10 Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.140322 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-vvnsb"] Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.149632 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.218625 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-vvnsb"] Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.226494 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-dns-svc\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.226597 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.226638 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.226699 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-config\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.226741 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f265w\" (UniqueName: \"kubernetes.io/projected/6de4174c-e8e2-4058-a35b-c277eced33f9-kube-api-access-f265w\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.226776 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.226817 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.328818 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.329341 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.329444 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-config\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.329487 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f265w\" (UniqueName: \"kubernetes.io/projected/6de4174c-e8e2-4058-a35b-c277eced33f9-kube-api-access-f265w\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.329515 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.329551 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.329605 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-dns-svc\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.330590 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-dns-svc\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.331429 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-config\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.331520 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.332317 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.332862 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.337658 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6de4174c-e8e2-4058-a35b-c277eced33f9-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.355658 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f265w\" (UniqueName: \"kubernetes.io/projected/6de4174c-e8e2-4058-a35b-c277eced33f9-kube-api-access-f265w\") pod \"dnsmasq-dns-55478c4467-vvnsb\" (UID: \"6de4174c-e8e2-4058-a35b-c277eced33f9\") " pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.488678 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.695351 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.741000 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-sb\") pod \"0888148b-128d-45c4-b6eb-20ce8d458e8c\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.741076 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-config\") pod \"0888148b-128d-45c4-b6eb-20ce8d458e8c\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.741201 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-nb\") pod \"0888148b-128d-45c4-b6eb-20ce8d458e8c\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.741271 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-swift-storage-0\") pod \"0888148b-128d-45c4-b6eb-20ce8d458e8c\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.741342 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-svc\") pod \"0888148b-128d-45c4-b6eb-20ce8d458e8c\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.741476 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc8l4\" (UniqueName: \"kubernetes.io/projected/0888148b-128d-45c4-b6eb-20ce8d458e8c-kube-api-access-wc8l4\") pod \"0888148b-128d-45c4-b6eb-20ce8d458e8c\" (UID: \"0888148b-128d-45c4-b6eb-20ce8d458e8c\") " Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.753063 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0888148b-128d-45c4-b6eb-20ce8d458e8c-kube-api-access-wc8l4" (OuterVolumeSpecName: "kube-api-access-wc8l4") pod "0888148b-128d-45c4-b6eb-20ce8d458e8c" (UID: "0888148b-128d-45c4-b6eb-20ce8d458e8c"). InnerVolumeSpecName "kube-api-access-wc8l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.806989 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0888148b-128d-45c4-b6eb-20ce8d458e8c" (UID: "0888148b-128d-45c4-b6eb-20ce8d458e8c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.820920 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0888148b-128d-45c4-b6eb-20ce8d458e8c" (UID: "0888148b-128d-45c4-b6eb-20ce8d458e8c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.826765 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0888148b-128d-45c4-b6eb-20ce8d458e8c" (UID: "0888148b-128d-45c4-b6eb-20ce8d458e8c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.830846 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-config" (OuterVolumeSpecName: "config") pod "0888148b-128d-45c4-b6eb-20ce8d458e8c" (UID: "0888148b-128d-45c4-b6eb-20ce8d458e8c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.834300 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0888148b-128d-45c4-b6eb-20ce8d458e8c" (UID: "0888148b-128d-45c4-b6eb-20ce8d458e8c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.844063 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.844115 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-config\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.844131 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.844140 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.844150 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0888148b-128d-45c4-b6eb-20ce8d458e8c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.844159 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc8l4\" (UniqueName: \"kubernetes.io/projected/0888148b-128d-45c4-b6eb-20ce8d458e8c-kube-api-access-wc8l4\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.937437 4876 generic.go:334] "Generic (PLEG): container finished" podID="0888148b-128d-45c4-b6eb-20ce8d458e8c" containerID="7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9" exitCode=0 Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.937496 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" event={"ID":"0888148b-128d-45c4-b6eb-20ce8d458e8c","Type":"ContainerDied","Data":"7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9"} Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.937538 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" event={"ID":"0888148b-128d-45c4-b6eb-20ce8d458e8c","Type":"ContainerDied","Data":"97c755411b53be6df78bb8b5ae422227f1ae28988f997de61f91409cf84cbedd"} Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.937560 4876 scope.go:117] "RemoveContainer" containerID="7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.937557 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-8dxbx" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.959528 4876 scope.go:117] "RemoveContainer" containerID="5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.996816 4876 scope.go:117] "RemoveContainer" containerID="7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9" Mar 13 08:03:00 crc kubenswrapper[4876]: E0313 08:03:00.997592 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9\": container with ID starting with 7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9 not found: ID does not exist" containerID="7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.997653 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9"} err="failed to get container status \"7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9\": rpc error: code = NotFound desc = could not find container \"7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9\": container with ID starting with 7a644338d263bd585c7fa6fc65158e7833cd004be0427f84e6e79ff9a5b239b9 not found: ID does not exist" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.997685 4876 scope.go:117] "RemoveContainer" containerID="5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6" Mar 13 08:03:00 crc kubenswrapper[4876]: E0313 08:03:00.998183 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6\": container with ID starting with 5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6 not found: ID does not exist" containerID="5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6" Mar 13 08:03:00 crc kubenswrapper[4876]: I0313 08:03:00.998207 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6"} err="failed to get container status \"5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6\": rpc error: code = NotFound desc = could not find container \"5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6\": container with ID starting with 5b8fc883a5efadad875788b351d0dc376ff98221d2b70106162936c780069ca6 not found: ID does not exist" Mar 13 08:03:01 crc kubenswrapper[4876]: I0313 08:03:01.000934 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8dxbx"] Mar 13 08:03:01 crc kubenswrapper[4876]: I0313 08:03:01.013143 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8dxbx"] Mar 13 08:03:01 crc kubenswrapper[4876]: I0313 08:03:01.052469 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0888148b-128d-45c4-b6eb-20ce8d458e8c" path="/var/lib/kubelet/pods/0888148b-128d-45c4-b6eb-20ce8d458e8c/volumes" Mar 13 08:03:01 crc kubenswrapper[4876]: I0313 08:03:01.098064 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-vvnsb"] Mar 13 08:03:01 crc kubenswrapper[4876]: E0313 08:03:01.585220 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6de4174c_e8e2_4058_a35b_c277eced33f9.slice/crio-conmon-04843e5bfdd172c1cc123f5427680b2f85ca71525ece974d26aff551fc7b8009.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6de4174c_e8e2_4058_a35b_c277eced33f9.slice/crio-04843e5bfdd172c1cc123f5427680b2f85ca71525ece974d26aff551fc7b8009.scope\": RecentStats: unable to find data in memory cache]" Mar 13 08:03:01 crc kubenswrapper[4876]: I0313 08:03:01.950678 4876 generic.go:334] "Generic (PLEG): container finished" podID="6de4174c-e8e2-4058-a35b-c277eced33f9" containerID="04843e5bfdd172c1cc123f5427680b2f85ca71525ece974d26aff551fc7b8009" exitCode=0 Mar 13 08:03:01 crc kubenswrapper[4876]: I0313 08:03:01.950778 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-vvnsb" event={"ID":"6de4174c-e8e2-4058-a35b-c277eced33f9","Type":"ContainerDied","Data":"04843e5bfdd172c1cc123f5427680b2f85ca71525ece974d26aff551fc7b8009"} Mar 13 08:03:01 crc kubenswrapper[4876]: I0313 08:03:01.951257 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-vvnsb" event={"ID":"6de4174c-e8e2-4058-a35b-c277eced33f9","Type":"ContainerStarted","Data":"85a14b195ac8d9d82053ba00e64145383b32747ec40b31cdad054b8bd0920709"} Mar 13 08:03:02 crc kubenswrapper[4876]: I0313 08:03:02.965382 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-vvnsb" event={"ID":"6de4174c-e8e2-4058-a35b-c277eced33f9","Type":"ContainerStarted","Data":"fc437ce78178cc81f7b62044b6174b944283a7da5b7d79c42cd62e5385988fe7"} Mar 13 08:03:02 crc kubenswrapper[4876]: I0313 08:03:02.965749 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:02 crc kubenswrapper[4876]: I0313 08:03:02.988856 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-vvnsb" podStartSLOduration=2.9888289070000003 podStartE2EDuration="2.988828907s" podCreationTimestamp="2026-03-13 08:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:03:02.985563185 +0000 UTC m=+1442.656342167" watchObservedRunningTime="2026-03-13 08:03:02.988828907 +0000 UTC m=+1442.659607889" Mar 13 08:03:10 crc kubenswrapper[4876]: I0313 08:03:10.490437 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-vvnsb" Mar 13 08:03:10 crc kubenswrapper[4876]: I0313 08:03:10.567867 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-xw8gl"] Mar 13 08:03:10 crc kubenswrapper[4876]: I0313 08:03:10.568202 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" podUID="bb7a1abc-aec3-4fab-82f6-9d305576db86" containerName="dnsmasq-dns" containerID="cri-o://50450a1e0da546a7f2ec5efe72a2a4ce3d0d9154e8b9b05e588df635f5db89ac" gracePeriod=10 Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.072220 4876 generic.go:334] "Generic (PLEG): container finished" podID="bb7a1abc-aec3-4fab-82f6-9d305576db86" containerID="50450a1e0da546a7f2ec5efe72a2a4ce3d0d9154e8b9b05e588df635f5db89ac" exitCode=0 Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.072329 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" event={"ID":"bb7a1abc-aec3-4fab-82f6-9d305576db86","Type":"ContainerDied","Data":"50450a1e0da546a7f2ec5efe72a2a4ce3d0d9154e8b9b05e588df635f5db89ac"} Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.072687 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" event={"ID":"bb7a1abc-aec3-4fab-82f6-9d305576db86","Type":"ContainerDied","Data":"f291fdd40c026e3e67b0fe039587b0e79e43a32c9eb6870268e7815bbe24c51d"} Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.072711 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f291fdd40c026e3e67b0fe039587b0e79e43a32c9eb6870268e7815bbe24c51d" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.087491 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.263157 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-openstack-edpm-ipam\") pod \"bb7a1abc-aec3-4fab-82f6-9d305576db86\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.263305 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-svc\") pod \"bb7a1abc-aec3-4fab-82f6-9d305576db86\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.263386 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-sb\") pod \"bb7a1abc-aec3-4fab-82f6-9d305576db86\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.263486 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-nb\") pod \"bb7a1abc-aec3-4fab-82f6-9d305576db86\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.263541 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwwqf\" (UniqueName: \"kubernetes.io/projected/bb7a1abc-aec3-4fab-82f6-9d305576db86-kube-api-access-pwwqf\") pod \"bb7a1abc-aec3-4fab-82f6-9d305576db86\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.263605 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-swift-storage-0\") pod \"bb7a1abc-aec3-4fab-82f6-9d305576db86\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.263673 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-config\") pod \"bb7a1abc-aec3-4fab-82f6-9d305576db86\" (UID: \"bb7a1abc-aec3-4fab-82f6-9d305576db86\") " Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.286581 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb7a1abc-aec3-4fab-82f6-9d305576db86-kube-api-access-pwwqf" (OuterVolumeSpecName: "kube-api-access-pwwqf") pod "bb7a1abc-aec3-4fab-82f6-9d305576db86" (UID: "bb7a1abc-aec3-4fab-82f6-9d305576db86"). InnerVolumeSpecName "kube-api-access-pwwqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.326487 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb7a1abc-aec3-4fab-82f6-9d305576db86" (UID: "bb7a1abc-aec3-4fab-82f6-9d305576db86"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.333723 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-config" (OuterVolumeSpecName: "config") pod "bb7a1abc-aec3-4fab-82f6-9d305576db86" (UID: "bb7a1abc-aec3-4fab-82f6-9d305576db86"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.329112 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb7a1abc-aec3-4fab-82f6-9d305576db86" (UID: "bb7a1abc-aec3-4fab-82f6-9d305576db86"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.335259 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb7a1abc-aec3-4fab-82f6-9d305576db86" (UID: "bb7a1abc-aec3-4fab-82f6-9d305576db86"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.339716 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb7a1abc-aec3-4fab-82f6-9d305576db86" (UID: "bb7a1abc-aec3-4fab-82f6-9d305576db86"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.355629 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "bb7a1abc-aec3-4fab-82f6-9d305576db86" (UID: "bb7a1abc-aec3-4fab-82f6-9d305576db86"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.366570 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.366623 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-config\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.366635 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.366651 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.366666 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.366677 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb7a1abc-aec3-4fab-82f6-9d305576db86-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:11 crc kubenswrapper[4876]: I0313 08:03:11.366688 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwwqf\" (UniqueName: \"kubernetes.io/projected/bb7a1abc-aec3-4fab-82f6-9d305576db86-kube-api-access-pwwqf\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:12 crc kubenswrapper[4876]: I0313 08:03:12.081109 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-xw8gl" Mar 13 08:03:12 crc kubenswrapper[4876]: I0313 08:03:12.119822 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-xw8gl"] Mar 13 08:03:12 crc kubenswrapper[4876]: I0313 08:03:12.131459 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-xw8gl"] Mar 13 08:03:13 crc kubenswrapper[4876]: I0313 08:03:13.049717 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb7a1abc-aec3-4fab-82f6-9d305576db86" path="/var/lib/kubelet/pods/bb7a1abc-aec3-4fab-82f6-9d305576db86/volumes" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.229542 4876 generic.go:334] "Generic (PLEG): container finished" podID="79aa8084-23ef-4730-948d-1b0d6ca25251" containerID="d8f014b4db9da3bcfc70ebac7e22b9b21887bfd1a56d20f8116edfc89e8c3b9d" exitCode=0 Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.229699 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"79aa8084-23ef-4730-948d-1b0d6ca25251","Type":"ContainerDied","Data":"d8f014b4db9da3bcfc70ebac7e22b9b21887bfd1a56d20f8116edfc89e8c3b9d"} Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.497610 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg"] Mar 13 08:03:23 crc kubenswrapper[4876]: E0313 08:03:23.498862 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0888148b-128d-45c4-b6eb-20ce8d458e8c" containerName="init" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.498966 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0888148b-128d-45c4-b6eb-20ce8d458e8c" containerName="init" Mar 13 08:03:23 crc kubenswrapper[4876]: E0313 08:03:23.499064 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0888148b-128d-45c4-b6eb-20ce8d458e8c" containerName="dnsmasq-dns" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.500189 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0888148b-128d-45c4-b6eb-20ce8d458e8c" containerName="dnsmasq-dns" Mar 13 08:03:23 crc kubenswrapper[4876]: E0313 08:03:23.500325 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7a1abc-aec3-4fab-82f6-9d305576db86" containerName="dnsmasq-dns" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.500396 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7a1abc-aec3-4fab-82f6-9d305576db86" containerName="dnsmasq-dns" Mar 13 08:03:23 crc kubenswrapper[4876]: E0313 08:03:23.500463 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7a1abc-aec3-4fab-82f6-9d305576db86" containerName="init" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.500666 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7a1abc-aec3-4fab-82f6-9d305576db86" containerName="init" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.501053 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0888148b-128d-45c4-b6eb-20ce8d458e8c" containerName="dnsmasq-dns" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.501124 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7a1abc-aec3-4fab-82f6-9d305576db86" containerName="dnsmasq-dns" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.502256 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.505757 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.506146 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.506520 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.508430 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.514113 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg"] Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.639097 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.639552 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4dgr\" (UniqueName: \"kubernetes.io/projected/61be635e-fed6-4b19-9095-cbaa3720b693-kube-api-access-b4dgr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.639681 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.639803 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.742106 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4dgr\" (UniqueName: \"kubernetes.io/projected/61be635e-fed6-4b19-9095-cbaa3720b693-kube-api-access-b4dgr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.742989 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.743836 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.744552 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.748586 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.749056 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.749696 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.763097 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4dgr\" (UniqueName: \"kubernetes.io/projected/61be635e-fed6-4b19-9095-cbaa3720b693-kube-api-access-b4dgr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:23 crc kubenswrapper[4876]: I0313 08:03:23.843310 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:24 crc kubenswrapper[4876]: I0313 08:03:24.242343 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"79aa8084-23ef-4730-948d-1b0d6ca25251","Type":"ContainerStarted","Data":"0ffb0f3d37877aca7bd80e8f9ac70e9b32b872bb636d8c47e2edfbd8cf3925fd"} Mar 13 08:03:24 crc kubenswrapper[4876]: I0313 08:03:24.242917 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 08:03:24 crc kubenswrapper[4876]: I0313 08:03:24.285032 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.28499544 podStartE2EDuration="37.28499544s" podCreationTimestamp="2026-03-13 08:02:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:03:24.27072081 +0000 UTC m=+1463.941499792" watchObservedRunningTime="2026-03-13 08:03:24.28499544 +0000 UTC m=+1463.955774422" Mar 13 08:03:24 crc kubenswrapper[4876]: I0313 08:03:24.492061 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:03:24 crc kubenswrapper[4876]: I0313 08:03:24.499267 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg"] Mar 13 08:03:25 crc kubenswrapper[4876]: I0313 08:03:25.257646 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" event={"ID":"61be635e-fed6-4b19-9095-cbaa3720b693","Type":"ContainerStarted","Data":"35baad4578c8f36dbc3a8c4287ea8a1067cb2fad3dc2e6334818e3c1295a201b"} Mar 13 08:03:26 crc kubenswrapper[4876]: I0313 08:03:26.273696 4876 generic.go:334] "Generic (PLEG): container finished" podID="147c88b9-db25-435e-9575-edb4efa342e7" containerID="c0a82e136f698070e4b8a7ca79ff71246b26ef7f3c3d2e4e7b12310a06bc08cd" exitCode=0 Mar 13 08:03:26 crc kubenswrapper[4876]: I0313 08:03:26.274064 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"147c88b9-db25-435e-9575-edb4efa342e7","Type":"ContainerDied","Data":"c0a82e136f698070e4b8a7ca79ff71246b26ef7f3c3d2e4e7b12310a06bc08cd"} Mar 13 08:03:28 crc kubenswrapper[4876]: I0313 08:03:28.299638 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"147c88b9-db25-435e-9575-edb4efa342e7","Type":"ContainerStarted","Data":"3d09113ae7a2e09725e317cc32210d59fccddf3e893e119a25de21486a390ac9"} Mar 13 08:03:28 crc kubenswrapper[4876]: I0313 08:03:28.300726 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:03:28 crc kubenswrapper[4876]: I0313 08:03:28.353225 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.353201108 podStartE2EDuration="38.353201108s" podCreationTimestamp="2026-03-13 08:02:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:03:28.340201813 +0000 UTC m=+1468.010980795" watchObservedRunningTime="2026-03-13 08:03:28.353201108 +0000 UTC m=+1468.023980080" Mar 13 08:03:35 crc kubenswrapper[4876]: I0313 08:03:35.373953 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" event={"ID":"61be635e-fed6-4b19-9095-cbaa3720b693","Type":"ContainerStarted","Data":"9711fac102a27d88613605e8b2699bd182c14e1f5b90cf5eadda9041512f024e"} Mar 13 08:03:35 crc kubenswrapper[4876]: I0313 08:03:35.403865 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" podStartSLOduration=1.865262463 podStartE2EDuration="12.40383619s" podCreationTimestamp="2026-03-13 08:03:23 +0000 UTC" firstStartedPulling="2026-03-13 08:03:24.491785266 +0000 UTC m=+1464.162564248" lastFinishedPulling="2026-03-13 08:03:35.030358993 +0000 UTC m=+1474.701137975" observedRunningTime="2026-03-13 08:03:35.40060727 +0000 UTC m=+1475.071386272" watchObservedRunningTime="2026-03-13 08:03:35.40383619 +0000 UTC m=+1475.074615172" Mar 13 08:03:38 crc kubenswrapper[4876]: I0313 08:03:38.486783 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 08:03:40 crc kubenswrapper[4876]: I0313 08:03:40.893646 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.557360 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6bkk8"] Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.559746 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.570117 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6bkk8"] Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.652527 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbm7r\" (UniqueName: \"kubernetes.io/projected/ba042ebf-8921-462a-a619-9a97ee404d05-kube-api-access-fbm7r\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.652764 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-catalog-content\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.653024 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-utilities\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.756062 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbm7r\" (UniqueName: \"kubernetes.io/projected/ba042ebf-8921-462a-a619-9a97ee404d05-kube-api-access-fbm7r\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.756173 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-catalog-content\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.756260 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-utilities\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.756766 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-catalog-content\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.756820 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-utilities\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.785551 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbm7r\" (UniqueName: \"kubernetes.io/projected/ba042ebf-8921-462a-a619-9a97ee404d05-kube-api-access-fbm7r\") pod \"redhat-operators-6bkk8\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:43 crc kubenswrapper[4876]: I0313 08:03:43.912179 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:44 crc kubenswrapper[4876]: I0313 08:03:44.450186 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6bkk8"] Mar 13 08:03:44 crc kubenswrapper[4876]: I0313 08:03:44.475970 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bkk8" event={"ID":"ba042ebf-8921-462a-a619-9a97ee404d05","Type":"ContainerStarted","Data":"9b37095928ca9085a9ced6d82f204bfb767a897ed989fb91f1b9d0d71d458587"} Mar 13 08:03:45 crc kubenswrapper[4876]: I0313 08:03:45.488885 4876 generic.go:334] "Generic (PLEG): container finished" podID="ba042ebf-8921-462a-a619-9a97ee404d05" containerID="1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8" exitCode=0 Mar 13 08:03:45 crc kubenswrapper[4876]: I0313 08:03:45.488993 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bkk8" event={"ID":"ba042ebf-8921-462a-a619-9a97ee404d05","Type":"ContainerDied","Data":"1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8"} Mar 13 08:03:46 crc kubenswrapper[4876]: I0313 08:03:46.503432 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bkk8" event={"ID":"ba042ebf-8921-462a-a619-9a97ee404d05","Type":"ContainerStarted","Data":"5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45"} Mar 13 08:03:49 crc kubenswrapper[4876]: I0313 08:03:49.541960 4876 generic.go:334] "Generic (PLEG): container finished" podID="61be635e-fed6-4b19-9095-cbaa3720b693" containerID="9711fac102a27d88613605e8b2699bd182c14e1f5b90cf5eadda9041512f024e" exitCode=0 Mar 13 08:03:49 crc kubenswrapper[4876]: I0313 08:03:49.542055 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" event={"ID":"61be635e-fed6-4b19-9095-cbaa3720b693","Type":"ContainerDied","Data":"9711fac102a27d88613605e8b2699bd182c14e1f5b90cf5eadda9041512f024e"} Mar 13 08:03:49 crc kubenswrapper[4876]: I0313 08:03:49.547120 4876 generic.go:334] "Generic (PLEG): container finished" podID="ba042ebf-8921-462a-a619-9a97ee404d05" containerID="5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45" exitCode=0 Mar 13 08:03:49 crc kubenswrapper[4876]: I0313 08:03:49.547180 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bkk8" event={"ID":"ba042ebf-8921-462a-a619-9a97ee404d05","Type":"ContainerDied","Data":"5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45"} Mar 13 08:03:50 crc kubenswrapper[4876]: I0313 08:03:50.562881 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bkk8" event={"ID":"ba042ebf-8921-462a-a619-9a97ee404d05","Type":"ContainerStarted","Data":"869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11"} Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.031999 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.059371 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6bkk8" podStartSLOduration=3.478508514 podStartE2EDuration="8.059339741s" podCreationTimestamp="2026-03-13 08:03:43 +0000 UTC" firstStartedPulling="2026-03-13 08:03:45.491343814 +0000 UTC m=+1485.162122796" lastFinishedPulling="2026-03-13 08:03:50.072175041 +0000 UTC m=+1489.742954023" observedRunningTime="2026-03-13 08:03:50.592533706 +0000 UTC m=+1490.263312688" watchObservedRunningTime="2026-03-13 08:03:51.059339741 +0000 UTC m=+1490.730118723" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.120481 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-ssh-key-openstack-edpm-ipam\") pod \"61be635e-fed6-4b19-9095-cbaa3720b693\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.120561 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4dgr\" (UniqueName: \"kubernetes.io/projected/61be635e-fed6-4b19-9095-cbaa3720b693-kube-api-access-b4dgr\") pod \"61be635e-fed6-4b19-9095-cbaa3720b693\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.120589 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-inventory\") pod \"61be635e-fed6-4b19-9095-cbaa3720b693\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.120839 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-repo-setup-combined-ca-bundle\") pod \"61be635e-fed6-4b19-9095-cbaa3720b693\" (UID: \"61be635e-fed6-4b19-9095-cbaa3720b693\") " Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.126788 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "61be635e-fed6-4b19-9095-cbaa3720b693" (UID: "61be635e-fed6-4b19-9095-cbaa3720b693"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.128523 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61be635e-fed6-4b19-9095-cbaa3720b693-kube-api-access-b4dgr" (OuterVolumeSpecName: "kube-api-access-b4dgr") pod "61be635e-fed6-4b19-9095-cbaa3720b693" (UID: "61be635e-fed6-4b19-9095-cbaa3720b693"). InnerVolumeSpecName "kube-api-access-b4dgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.152867 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "61be635e-fed6-4b19-9095-cbaa3720b693" (UID: "61be635e-fed6-4b19-9095-cbaa3720b693"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.158346 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-inventory" (OuterVolumeSpecName: "inventory") pod "61be635e-fed6-4b19-9095-cbaa3720b693" (UID: "61be635e-fed6-4b19-9095-cbaa3720b693"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.224560 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.224615 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.224629 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4dgr\" (UniqueName: \"kubernetes.io/projected/61be635e-fed6-4b19-9095-cbaa3720b693-kube-api-access-b4dgr\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.224644 4876 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61be635e-fed6-4b19-9095-cbaa3720b693-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.575979 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" event={"ID":"61be635e-fed6-4b19-9095-cbaa3720b693","Type":"ContainerDied","Data":"35baad4578c8f36dbc3a8c4287ea8a1067cb2fad3dc2e6334818e3c1295a201b"} Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.576044 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35baad4578c8f36dbc3a8c4287ea8a1067cb2fad3dc2e6334818e3c1295a201b" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.577389 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.664319 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh"] Mar 13 08:03:51 crc kubenswrapper[4876]: E0313 08:03:51.664881 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61be635e-fed6-4b19-9095-cbaa3720b693" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.664909 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="61be635e-fed6-4b19-9095-cbaa3720b693" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.665139 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="61be635e-fed6-4b19-9095-cbaa3720b693" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.665908 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.668830 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.668921 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.669130 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.670924 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.681484 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh"] Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.735727 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g4r6\" (UniqueName: \"kubernetes.io/projected/1f20780c-14f1-4165-95a5-15b604018e4b-kube-api-access-4g4r6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.735791 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.735852 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.837611 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g4r6\" (UniqueName: \"kubernetes.io/projected/1f20780c-14f1-4165-95a5-15b604018e4b-kube-api-access-4g4r6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.838043 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.838133 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.845039 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.855968 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.860025 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g4r6\" (UniqueName: \"kubernetes.io/projected/1f20780c-14f1-4165-95a5-15b604018e4b-kube-api-access-4g4r6\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cm4qh\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:51 crc kubenswrapper[4876]: I0313 08:03:51.984200 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:52 crc kubenswrapper[4876]: I0313 08:03:52.682223 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh"] Mar 13 08:03:53 crc kubenswrapper[4876]: I0313 08:03:53.598108 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" event={"ID":"1f20780c-14f1-4165-95a5-15b604018e4b","Type":"ContainerStarted","Data":"3b259c9a47508f27c5ee89b7579f134a0ed4161cf4fc8a6bc19a12d485c1701d"} Mar 13 08:03:53 crc kubenswrapper[4876]: I0313 08:03:53.912479 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:53 crc kubenswrapper[4876]: I0313 08:03:53.913982 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:03:54 crc kubenswrapper[4876]: I0313 08:03:54.611757 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" event={"ID":"1f20780c-14f1-4165-95a5-15b604018e4b","Type":"ContainerStarted","Data":"ae840cce8ba8121560d01107e8896d79809f359f77fe378048d2ecbd33923c5e"} Mar 13 08:03:54 crc kubenswrapper[4876]: I0313 08:03:54.632498 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" podStartSLOduration=2.119649729 podStartE2EDuration="3.632466962s" podCreationTimestamp="2026-03-13 08:03:51 +0000 UTC" firstStartedPulling="2026-03-13 08:03:52.687746673 +0000 UTC m=+1492.358525655" lastFinishedPulling="2026-03-13 08:03:54.200563906 +0000 UTC m=+1493.871342888" observedRunningTime="2026-03-13 08:03:54.628016268 +0000 UTC m=+1494.298795260" watchObservedRunningTime="2026-03-13 08:03:54.632466962 +0000 UTC m=+1494.303245944" Mar 13 08:03:54 crc kubenswrapper[4876]: I0313 08:03:54.963117 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6bkk8" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="registry-server" probeResult="failure" output=< Mar 13 08:03:54 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:03:54 crc kubenswrapper[4876]: > Mar 13 08:03:57 crc kubenswrapper[4876]: I0313 08:03:57.652322 4876 generic.go:334] "Generic (PLEG): container finished" podID="1f20780c-14f1-4165-95a5-15b604018e4b" containerID="ae840cce8ba8121560d01107e8896d79809f359f77fe378048d2ecbd33923c5e" exitCode=0 Mar 13 08:03:57 crc kubenswrapper[4876]: I0313 08:03:57.652427 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" event={"ID":"1f20780c-14f1-4165-95a5-15b604018e4b","Type":"ContainerDied","Data":"ae840cce8ba8121560d01107e8896d79809f359f77fe378048d2ecbd33923c5e"} Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.157191 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.220379 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-ssh-key-openstack-edpm-ipam\") pod \"1f20780c-14f1-4165-95a5-15b604018e4b\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.220588 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-inventory\") pod \"1f20780c-14f1-4165-95a5-15b604018e4b\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.220675 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g4r6\" (UniqueName: \"kubernetes.io/projected/1f20780c-14f1-4165-95a5-15b604018e4b-kube-api-access-4g4r6\") pod \"1f20780c-14f1-4165-95a5-15b604018e4b\" (UID: \"1f20780c-14f1-4165-95a5-15b604018e4b\") " Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.227816 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f20780c-14f1-4165-95a5-15b604018e4b-kube-api-access-4g4r6" (OuterVolumeSpecName: "kube-api-access-4g4r6") pod "1f20780c-14f1-4165-95a5-15b604018e4b" (UID: "1f20780c-14f1-4165-95a5-15b604018e4b"). InnerVolumeSpecName "kube-api-access-4g4r6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.253542 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1f20780c-14f1-4165-95a5-15b604018e4b" (UID: "1f20780c-14f1-4165-95a5-15b604018e4b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.254759 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-inventory" (OuterVolumeSpecName: "inventory") pod "1f20780c-14f1-4165-95a5-15b604018e4b" (UID: "1f20780c-14f1-4165-95a5-15b604018e4b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.323805 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.323887 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f20780c-14f1-4165-95a5-15b604018e4b-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.323897 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g4r6\" (UniqueName: \"kubernetes.io/projected/1f20780c-14f1-4165-95a5-15b604018e4b-kube-api-access-4g4r6\") on node \"crc\" DevicePath \"\"" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.684594 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" event={"ID":"1f20780c-14f1-4165-95a5-15b604018e4b","Type":"ContainerDied","Data":"3b259c9a47508f27c5ee89b7579f134a0ed4161cf4fc8a6bc19a12d485c1701d"} Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.684647 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b259c9a47508f27c5ee89b7579f134a0ed4161cf4fc8a6bc19a12d485c1701d" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.684671 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cm4qh" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.769659 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z"] Mar 13 08:03:59 crc kubenswrapper[4876]: E0313 08:03:59.770211 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f20780c-14f1-4165-95a5-15b604018e4b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.770255 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f20780c-14f1-4165-95a5-15b604018e4b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.770473 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f20780c-14f1-4165-95a5-15b604018e4b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.771283 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.777206 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.777597 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.777976 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.779538 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.787092 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z"] Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.831927 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.832321 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.832393 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt6cq\" (UniqueName: \"kubernetes.io/projected/0808c5a4-53f5-43e5-af39-bf78530e1d71-kube-api-access-bt6cq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.832449 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.934249 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.934377 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.934418 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.934505 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt6cq\" (UniqueName: \"kubernetes.io/projected/0808c5a4-53f5-43e5-af39-bf78530e1d71-kube-api-access-bt6cq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.939525 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.940042 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.940215 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:03:59 crc kubenswrapper[4876]: I0313 08:03:59.951205 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt6cq\" (UniqueName: \"kubernetes.io/projected/0808c5a4-53f5-43e5-af39-bf78530e1d71-kube-api-access-bt6cq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.091776 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.144424 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556484-mlmsn"] Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.146111 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556484-mlmsn" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.148662 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.149140 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.154866 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.155552 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556484-mlmsn"] Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.240493 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzv2k\" (UniqueName: \"kubernetes.io/projected/7fe1c50d-4a24-479d-9d08-780fc1c0e0a7-kube-api-access-wzv2k\") pod \"auto-csr-approver-29556484-mlmsn\" (UID: \"7fe1c50d-4a24-479d-9d08-780fc1c0e0a7\") " pod="openshift-infra/auto-csr-approver-29556484-mlmsn" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.343190 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzv2k\" (UniqueName: \"kubernetes.io/projected/7fe1c50d-4a24-479d-9d08-780fc1c0e0a7-kube-api-access-wzv2k\") pod \"auto-csr-approver-29556484-mlmsn\" (UID: \"7fe1c50d-4a24-479d-9d08-780fc1c0e0a7\") " pod="openshift-infra/auto-csr-approver-29556484-mlmsn" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.364368 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzv2k\" (UniqueName: \"kubernetes.io/projected/7fe1c50d-4a24-479d-9d08-780fc1c0e0a7-kube-api-access-wzv2k\") pod \"auto-csr-approver-29556484-mlmsn\" (UID: \"7fe1c50d-4a24-479d-9d08-780fc1c0e0a7\") " pod="openshift-infra/auto-csr-approver-29556484-mlmsn" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.560901 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556484-mlmsn" Mar 13 08:04:00 crc kubenswrapper[4876]: I0313 08:04:00.727031 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z"] Mar 13 08:04:01 crc kubenswrapper[4876]: I0313 08:04:01.088964 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556484-mlmsn"] Mar 13 08:04:01 crc kubenswrapper[4876]: I0313 08:04:01.284553 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:04:01 crc kubenswrapper[4876]: I0313 08:04:01.746989 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" event={"ID":"0808c5a4-53f5-43e5-af39-bf78530e1d71","Type":"ContainerStarted","Data":"474f8ca4ddf1c9df45575afff5bb96e4d8f7a63f3f522f2d40842f4ac083fe97"} Mar 13 08:04:01 crc kubenswrapper[4876]: I0313 08:04:01.747559 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" event={"ID":"0808c5a4-53f5-43e5-af39-bf78530e1d71","Type":"ContainerStarted","Data":"cba143fd90d7931f15ad699d1f38a8462866f887d1069d5aae7d6990ad8e7c4b"} Mar 13 08:04:01 crc kubenswrapper[4876]: I0313 08:04:01.762324 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556484-mlmsn" event={"ID":"7fe1c50d-4a24-479d-9d08-780fc1c0e0a7","Type":"ContainerStarted","Data":"4fce84bfaed6d7c5a72e6c8d742a44d0a1c063086b45776e1fbb2111ee70dc08"} Mar 13 08:04:02 crc kubenswrapper[4876]: I0313 08:04:02.775497 4876 generic.go:334] "Generic (PLEG): container finished" podID="7fe1c50d-4a24-479d-9d08-780fc1c0e0a7" containerID="ef20163b689236111fb8e6fc380a33a061900a45cc3f2e63ffa286b5d0215ea7" exitCode=0 Mar 13 08:04:02 crc kubenswrapper[4876]: I0313 08:04:02.775608 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556484-mlmsn" event={"ID":"7fe1c50d-4a24-479d-9d08-780fc1c0e0a7","Type":"ContainerDied","Data":"ef20163b689236111fb8e6fc380a33a061900a45cc3f2e63ffa286b5d0215ea7"} Mar 13 08:04:02 crc kubenswrapper[4876]: I0313 08:04:02.799215 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" podStartSLOduration=3.251625329 podStartE2EDuration="3.799189817s" podCreationTimestamp="2026-03-13 08:03:59 +0000 UTC" firstStartedPulling="2026-03-13 08:04:00.733432436 +0000 UTC m=+1500.404211418" lastFinishedPulling="2026-03-13 08:04:01.280996914 +0000 UTC m=+1500.951775906" observedRunningTime="2026-03-13 08:04:01.791983747 +0000 UTC m=+1501.462762739" watchObservedRunningTime="2026-03-13 08:04:02.799189817 +0000 UTC m=+1502.469968799" Mar 13 08:04:03 crc kubenswrapper[4876]: I0313 08:04:03.976363 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.060272 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.223195 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6bkk8"] Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.253891 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556484-mlmsn" Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.338147 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzv2k\" (UniqueName: \"kubernetes.io/projected/7fe1c50d-4a24-479d-9d08-780fc1c0e0a7-kube-api-access-wzv2k\") pod \"7fe1c50d-4a24-479d-9d08-780fc1c0e0a7\" (UID: \"7fe1c50d-4a24-479d-9d08-780fc1c0e0a7\") " Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.354826 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe1c50d-4a24-479d-9d08-780fc1c0e0a7-kube-api-access-wzv2k" (OuterVolumeSpecName: "kube-api-access-wzv2k") pod "7fe1c50d-4a24-479d-9d08-780fc1c0e0a7" (UID: "7fe1c50d-4a24-479d-9d08-780fc1c0e0a7"). InnerVolumeSpecName "kube-api-access-wzv2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.440886 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzv2k\" (UniqueName: \"kubernetes.io/projected/7fe1c50d-4a24-479d-9d08-780fc1c0e0a7-kube-api-access-wzv2k\") on node \"crc\" DevicePath \"\"" Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.797969 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556484-mlmsn" event={"ID":"7fe1c50d-4a24-479d-9d08-780fc1c0e0a7","Type":"ContainerDied","Data":"4fce84bfaed6d7c5a72e6c8d742a44d0a1c063086b45776e1fbb2111ee70dc08"} Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.798028 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fce84bfaed6d7c5a72e6c8d742a44d0a1c063086b45776e1fbb2111ee70dc08" Mar 13 08:04:04 crc kubenswrapper[4876]: I0313 08:04:04.797992 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556484-mlmsn" Mar 13 08:04:05 crc kubenswrapper[4876]: I0313 08:04:05.343789 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556478-vspx7"] Mar 13 08:04:05 crc kubenswrapper[4876]: I0313 08:04:05.354537 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556478-vspx7"] Mar 13 08:04:05 crc kubenswrapper[4876]: I0313 08:04:05.807896 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6bkk8" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="registry-server" containerID="cri-o://869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11" gracePeriod=2 Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.294578 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.387347 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbm7r\" (UniqueName: \"kubernetes.io/projected/ba042ebf-8921-462a-a619-9a97ee404d05-kube-api-access-fbm7r\") pod \"ba042ebf-8921-462a-a619-9a97ee404d05\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.387483 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-utilities\") pod \"ba042ebf-8921-462a-a619-9a97ee404d05\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.388227 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-catalog-content\") pod \"ba042ebf-8921-462a-a619-9a97ee404d05\" (UID: \"ba042ebf-8921-462a-a619-9a97ee404d05\") " Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.388623 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-utilities" (OuterVolumeSpecName: "utilities") pod "ba042ebf-8921-462a-a619-9a97ee404d05" (UID: "ba042ebf-8921-462a-a619-9a97ee404d05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.389633 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.395383 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba042ebf-8921-462a-a619-9a97ee404d05-kube-api-access-fbm7r" (OuterVolumeSpecName: "kube-api-access-fbm7r") pod "ba042ebf-8921-462a-a619-9a97ee404d05" (UID: "ba042ebf-8921-462a-a619-9a97ee404d05"). InnerVolumeSpecName "kube-api-access-fbm7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.414224 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7b9c5d6d4c-7lxwx" podUID="62fd9eb3-6af0-4bc9-bc57-65c8b03eb467" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.492067 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbm7r\" (UniqueName: \"kubernetes.io/projected/ba042ebf-8921-462a-a619-9a97ee404d05-kube-api-access-fbm7r\") on node \"crc\" DevicePath \"\"" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.548626 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba042ebf-8921-462a-a619-9a97ee404d05" (UID: "ba042ebf-8921-462a-a619-9a97ee404d05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.594587 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba042ebf-8921-462a-a619-9a97ee404d05-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.823937 4876 generic.go:334] "Generic (PLEG): container finished" podID="ba042ebf-8921-462a-a619-9a97ee404d05" containerID="869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11" exitCode=0 Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.824002 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bkk8" event={"ID":"ba042ebf-8921-462a-a619-9a97ee404d05","Type":"ContainerDied","Data":"869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11"} Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.824043 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bkk8" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.824071 4876 scope.go:117] "RemoveContainer" containerID="869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.824053 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bkk8" event={"ID":"ba042ebf-8921-462a-a619-9a97ee404d05","Type":"ContainerDied","Data":"9b37095928ca9085a9ced6d82f204bfb767a897ed989fb91f1b9d0d71d458587"} Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.847985 4876 scope.go:117] "RemoveContainer" containerID="5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.894341 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6bkk8"] Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.902881 4876 scope.go:117] "RemoveContainer" containerID="1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.912427 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6bkk8"] Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.942406 4876 scope.go:117] "RemoveContainer" containerID="869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11" Mar 13 08:04:06 crc kubenswrapper[4876]: E0313 08:04:06.943004 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11\": container with ID starting with 869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11 not found: ID does not exist" containerID="869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.943055 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11"} err="failed to get container status \"869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11\": rpc error: code = NotFound desc = could not find container \"869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11\": container with ID starting with 869cf7ffe03f6673b6d5ba0862ad94973565ff4bb75d11530f89663afe6f8f11 not found: ID does not exist" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.943094 4876 scope.go:117] "RemoveContainer" containerID="5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45" Mar 13 08:04:06 crc kubenswrapper[4876]: E0313 08:04:06.943492 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45\": container with ID starting with 5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45 not found: ID does not exist" containerID="5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.943532 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45"} err="failed to get container status \"5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45\": rpc error: code = NotFound desc = could not find container \"5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45\": container with ID starting with 5dc096ee352991ebbee52f73d0967498c7b3ad1ba0682d5f6ad85f6f7b4dbb45 not found: ID does not exist" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.943555 4876 scope.go:117] "RemoveContainer" containerID="1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8" Mar 13 08:04:06 crc kubenswrapper[4876]: E0313 08:04:06.943780 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8\": container with ID starting with 1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8 not found: ID does not exist" containerID="1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8" Mar 13 08:04:06 crc kubenswrapper[4876]: I0313 08:04:06.943803 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8"} err="failed to get container status \"1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8\": rpc error: code = NotFound desc = could not find container \"1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8\": container with ID starting with 1ec4ddfe0e7b2765c4d719a877a26650537ed354054f26c9cba0847ad8f114d8 not found: ID does not exist" Mar 13 08:04:07 crc kubenswrapper[4876]: I0313 08:04:07.051835 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99cd306b-1011-48b5-8dfd-b70f6e28f057" path="/var/lib/kubelet/pods/99cd306b-1011-48b5-8dfd-b70f6e28f057/volumes" Mar 13 08:04:07 crc kubenswrapper[4876]: I0313 08:04:07.052734 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" path="/var/lib/kubelet/pods/ba042ebf-8921-462a-a619-9a97ee404d05/volumes" Mar 13 08:04:19 crc kubenswrapper[4876]: I0313 08:04:19.153000 4876 scope.go:117] "RemoveContainer" containerID="0054e18eed70fb08e7510798b681c205692d43652c7d8a318ba96752774ce19c" Mar 13 08:04:19 crc kubenswrapper[4876]: I0313 08:04:19.188794 4876 scope.go:117] "RemoveContainer" containerID="d7e72f20ba8a3569864c6741c938b22f3bb790db5a73b9bb43f7e825d3a9aa14" Mar 13 08:04:19 crc kubenswrapper[4876]: I0313 08:04:19.248651 4876 scope.go:117] "RemoveContainer" containerID="d9c45b3607ae9397dd7932a99ff50e2d3ef797c6a4f714c8947b7859a18a262e" Mar 13 08:04:54 crc kubenswrapper[4876]: I0313 08:04:54.607437 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:04:54 crc kubenswrapper[4876]: I0313 08:04:54.608814 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:05:19 crc kubenswrapper[4876]: I0313 08:05:19.397200 4876 scope.go:117] "RemoveContainer" containerID="78a12f079c86c03a1434cab042a6e40ad549af96a3e8a139ac0f74fb201bedce" Mar 13 08:05:19 crc kubenswrapper[4876]: I0313 08:05:19.436609 4876 scope.go:117] "RemoveContainer" containerID="d21cc25bd945c08dee67e81d7642e373e6e49cc67322be5d3e60e5c28f6d5bf5" Mar 13 08:05:19 crc kubenswrapper[4876]: I0313 08:05:19.462356 4876 scope.go:117] "RemoveContainer" containerID="2e99ade3cdbc5460384ad6c59b414c3d139f3f8b1523d2abf7cd47152566ea2a" Mar 13 08:05:19 crc kubenswrapper[4876]: I0313 08:05:19.492892 4876 scope.go:117] "RemoveContainer" containerID="80661cfa70c99dd7b3d101dbcd3d9f87ed202b141eb93e53b24e735b71649c7c" Mar 13 08:05:19 crc kubenswrapper[4876]: I0313 08:05:19.515979 4876 scope.go:117] "RemoveContainer" containerID="a82353ca03963734c9e8d399b80dc22325e344e78ee20af877d27c28fbfe3390" Mar 13 08:05:19 crc kubenswrapper[4876]: I0313 08:05:19.542604 4876 scope.go:117] "RemoveContainer" containerID="d920151cf960cd9011266a42d377ced5f63cc0564624c236cf81e9406ae2903b" Mar 13 08:05:19 crc kubenswrapper[4876]: I0313 08:05:19.616890 4876 scope.go:117] "RemoveContainer" containerID="50e23ad80caa9b0a975b0d31c0f0ca3c2d7d59e5b837f13c9b725b68dbce8a4e" Mar 13 08:05:24 crc kubenswrapper[4876]: I0313 08:05:24.608008 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:05:24 crc kubenswrapper[4876]: I0313 08:05:24.608691 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.586075 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9lsfg"] Mar 13 08:05:46 crc kubenswrapper[4876]: E0313 08:05:46.587299 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="extract-utilities" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.587318 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="extract-utilities" Mar 13 08:05:46 crc kubenswrapper[4876]: E0313 08:05:46.587341 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="registry-server" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.587348 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="registry-server" Mar 13 08:05:46 crc kubenswrapper[4876]: E0313 08:05:46.587376 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="extract-content" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.587386 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="extract-content" Mar 13 08:05:46 crc kubenswrapper[4876]: E0313 08:05:46.587416 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe1c50d-4a24-479d-9d08-780fc1c0e0a7" containerName="oc" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.587424 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe1c50d-4a24-479d-9d08-780fc1c0e0a7" containerName="oc" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.587633 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba042ebf-8921-462a-a619-9a97ee404d05" containerName="registry-server" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.587657 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe1c50d-4a24-479d-9d08-780fc1c0e0a7" containerName="oc" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.589612 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.606340 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lsfg"] Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.754613 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-catalog-content\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.754876 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-utilities\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.755148 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpcsm\" (UniqueName: \"kubernetes.io/projected/643303f7-da91-41de-8fe6-407a7794de70-kube-api-access-wpcsm\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.857175 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpcsm\" (UniqueName: \"kubernetes.io/projected/643303f7-da91-41de-8fe6-407a7794de70-kube-api-access-wpcsm\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.857497 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-catalog-content\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.857562 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-utilities\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.858209 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-utilities\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.858335 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-catalog-content\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.888508 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpcsm\" (UniqueName: \"kubernetes.io/projected/643303f7-da91-41de-8fe6-407a7794de70-kube-api-access-wpcsm\") pod \"redhat-marketplace-9lsfg\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:46 crc kubenswrapper[4876]: I0313 08:05:46.921717 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:47 crc kubenswrapper[4876]: I0313 08:05:47.456974 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lsfg"] Mar 13 08:05:47 crc kubenswrapper[4876]: I0313 08:05:47.964179 4876 generic.go:334] "Generic (PLEG): container finished" podID="643303f7-da91-41de-8fe6-407a7794de70" containerID="9c805176a408f3ddb2df890a616db47edb2d137dc31bb29a5baacfe27370b7ab" exitCode=0 Mar 13 08:05:47 crc kubenswrapper[4876]: I0313 08:05:47.964231 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lsfg" event={"ID":"643303f7-da91-41de-8fe6-407a7794de70","Type":"ContainerDied","Data":"9c805176a408f3ddb2df890a616db47edb2d137dc31bb29a5baacfe27370b7ab"} Mar 13 08:05:47 crc kubenswrapper[4876]: I0313 08:05:47.964291 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lsfg" event={"ID":"643303f7-da91-41de-8fe6-407a7794de70","Type":"ContainerStarted","Data":"b2d1bb1b52fc5d03249a3b8709e72f6ee881437db2d2d327e6d61e5ed3be4953"} Mar 13 08:05:48 crc kubenswrapper[4876]: I0313 08:05:48.977178 4876 generic.go:334] "Generic (PLEG): container finished" podID="643303f7-da91-41de-8fe6-407a7794de70" containerID="fa619c02492710df3118fa5e92ce5fed9bbddf8b3dc470b1fd01fc684d4f2942" exitCode=0 Mar 13 08:05:48 crc kubenswrapper[4876]: I0313 08:05:48.977283 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lsfg" event={"ID":"643303f7-da91-41de-8fe6-407a7794de70","Type":"ContainerDied","Data":"fa619c02492710df3118fa5e92ce5fed9bbddf8b3dc470b1fd01fc684d4f2942"} Mar 13 08:05:49 crc kubenswrapper[4876]: I0313 08:05:49.992624 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lsfg" event={"ID":"643303f7-da91-41de-8fe6-407a7794de70","Type":"ContainerStarted","Data":"f56c3f230c2f11f80c75e9b403149d9aefb10f7aaefb8d78e497bbb88bb623a6"} Mar 13 08:05:50 crc kubenswrapper[4876]: I0313 08:05:50.015343 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9lsfg" podStartSLOduration=2.527280773 podStartE2EDuration="4.015316572s" podCreationTimestamp="2026-03-13 08:05:46 +0000 UTC" firstStartedPulling="2026-03-13 08:05:47.967937156 +0000 UTC m=+1607.638716138" lastFinishedPulling="2026-03-13 08:05:49.455972955 +0000 UTC m=+1609.126751937" observedRunningTime="2026-03-13 08:05:50.009039357 +0000 UTC m=+1609.679818349" watchObservedRunningTime="2026-03-13 08:05:50.015316572 +0000 UTC m=+1609.686095554" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.205020 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zqccs"] Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.208491 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.221997 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqccs"] Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.311189 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv5z8\" (UniqueName: \"kubernetes.io/projected/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-kube-api-access-vv5z8\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.311325 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-utilities\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.311506 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-catalog-content\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.413288 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv5z8\" (UniqueName: \"kubernetes.io/projected/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-kube-api-access-vv5z8\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.413353 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-utilities\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.413478 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-catalog-content\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.413969 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-utilities\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.414129 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-catalog-content\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.437373 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv5z8\" (UniqueName: \"kubernetes.io/projected/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-kube-api-access-vv5z8\") pod \"community-operators-zqccs\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.550759 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.607543 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.607635 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.607706 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.608848 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:05:54 crc kubenswrapper[4876]: I0313 08:05:54.608915 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" gracePeriod=600 Mar 13 08:05:54 crc kubenswrapper[4876]: E0313 08:05:54.760705 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:05:55 crc kubenswrapper[4876]: I0313 08:05:55.050399 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" exitCode=0 Mar 13 08:05:55 crc kubenswrapper[4876]: I0313 08:05:55.050468 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2"} Mar 13 08:05:55 crc kubenswrapper[4876]: I0313 08:05:55.050877 4876 scope.go:117] "RemoveContainer" containerID="ab9fcd397cad6841be9000d6381846b64e56e1d04c705091fe8c6b9884bbef7d" Mar 13 08:05:55 crc kubenswrapper[4876]: I0313 08:05:55.051702 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:05:55 crc kubenswrapper[4876]: E0313 08:05:55.052014 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:05:55 crc kubenswrapper[4876]: I0313 08:05:55.126023 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqccs"] Mar 13 08:05:56 crc kubenswrapper[4876]: I0313 08:05:56.071739 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerID="58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda" exitCode=0 Mar 13 08:05:56 crc kubenswrapper[4876]: I0313 08:05:56.071814 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqccs" event={"ID":"0a123dd8-46d2-4bf9-92a7-34151f54c0e1","Type":"ContainerDied","Data":"58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda"} Mar 13 08:05:56 crc kubenswrapper[4876]: I0313 08:05:56.078500 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqccs" event={"ID":"0a123dd8-46d2-4bf9-92a7-34151f54c0e1","Type":"ContainerStarted","Data":"40ce537937db57e137ac20976f94440080877e972ddf0eab0c158b25188196a0"} Mar 13 08:05:56 crc kubenswrapper[4876]: I0313 08:05:56.922709 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:56 crc kubenswrapper[4876]: I0313 08:05:56.923155 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:56 crc kubenswrapper[4876]: I0313 08:05:56.971202 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:57 crc kubenswrapper[4876]: I0313 08:05:57.144801 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:05:58 crc kubenswrapper[4876]: I0313 08:05:58.112956 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerID="3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809" exitCode=0 Mar 13 08:05:58 crc kubenswrapper[4876]: I0313 08:05:58.114610 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqccs" event={"ID":"0a123dd8-46d2-4bf9-92a7-34151f54c0e1","Type":"ContainerDied","Data":"3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809"} Mar 13 08:05:58 crc kubenswrapper[4876]: I0313 08:05:58.985191 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lsfg"] Mar 13 08:05:59 crc kubenswrapper[4876]: I0313 08:05:59.209799 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9lsfg" podUID="643303f7-da91-41de-8fe6-407a7794de70" containerName="registry-server" containerID="cri-o://f56c3f230c2f11f80c75e9b403149d9aefb10f7aaefb8d78e497bbb88bb623a6" gracePeriod=2 Mar 13 08:05:59 crc kubenswrapper[4876]: I0313 08:05:59.210343 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqccs" event={"ID":"0a123dd8-46d2-4bf9-92a7-34151f54c0e1","Type":"ContainerStarted","Data":"3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0"} Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.152667 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zqccs" podStartSLOduration=3.694960757 podStartE2EDuration="6.152640283s" podCreationTimestamp="2026-03-13 08:05:54 +0000 UTC" firstStartedPulling="2026-03-13 08:05:56.073974342 +0000 UTC m=+1615.744753334" lastFinishedPulling="2026-03-13 08:05:58.531653878 +0000 UTC m=+1618.202432860" observedRunningTime="2026-03-13 08:05:59.255716443 +0000 UTC m=+1618.926495425" watchObservedRunningTime="2026-03-13 08:06:00.152640283 +0000 UTC m=+1619.823419265" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.172951 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556486-zprrc"] Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.174719 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556486-zprrc" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.177390 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.177479 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.177607 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.193414 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556486-zprrc"] Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.222257 4876 generic.go:334] "Generic (PLEG): container finished" podID="643303f7-da91-41de-8fe6-407a7794de70" containerID="f56c3f230c2f11f80c75e9b403149d9aefb10f7aaefb8d78e497bbb88bb623a6" exitCode=0 Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.222280 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lsfg" event={"ID":"643303f7-da91-41de-8fe6-407a7794de70","Type":"ContainerDied","Data":"f56c3f230c2f11f80c75e9b403149d9aefb10f7aaefb8d78e497bbb88bb623a6"} Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.222349 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9lsfg" event={"ID":"643303f7-da91-41de-8fe6-407a7794de70","Type":"ContainerDied","Data":"b2d1bb1b52fc5d03249a3b8709e72f6ee881437db2d2d327e6d61e5ed3be4953"} Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.222369 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2d1bb1b52fc5d03249a3b8709e72f6ee881437db2d2d327e6d61e5ed3be4953" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.247406 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.293657 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqqnf\" (UniqueName: \"kubernetes.io/projected/770aa8c1-c4df-4219-bad8-f77761708cb7-kube-api-access-nqqnf\") pod \"auto-csr-approver-29556486-zprrc\" (UID: \"770aa8c1-c4df-4219-bad8-f77761708cb7\") " pod="openshift-infra/auto-csr-approver-29556486-zprrc" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.395089 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-catalog-content\") pod \"643303f7-da91-41de-8fe6-407a7794de70\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.395218 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-utilities\") pod \"643303f7-da91-41de-8fe6-407a7794de70\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.395291 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpcsm\" (UniqueName: \"kubernetes.io/projected/643303f7-da91-41de-8fe6-407a7794de70-kube-api-access-wpcsm\") pod \"643303f7-da91-41de-8fe6-407a7794de70\" (UID: \"643303f7-da91-41de-8fe6-407a7794de70\") " Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.395845 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqqnf\" (UniqueName: \"kubernetes.io/projected/770aa8c1-c4df-4219-bad8-f77761708cb7-kube-api-access-nqqnf\") pod \"auto-csr-approver-29556486-zprrc\" (UID: \"770aa8c1-c4df-4219-bad8-f77761708cb7\") " pod="openshift-infra/auto-csr-approver-29556486-zprrc" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.397305 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-utilities" (OuterVolumeSpecName: "utilities") pod "643303f7-da91-41de-8fe6-407a7794de70" (UID: "643303f7-da91-41de-8fe6-407a7794de70"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.403033 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/643303f7-da91-41de-8fe6-407a7794de70-kube-api-access-wpcsm" (OuterVolumeSpecName: "kube-api-access-wpcsm") pod "643303f7-da91-41de-8fe6-407a7794de70" (UID: "643303f7-da91-41de-8fe6-407a7794de70"). InnerVolumeSpecName "kube-api-access-wpcsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.416589 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqqnf\" (UniqueName: \"kubernetes.io/projected/770aa8c1-c4df-4219-bad8-f77761708cb7-kube-api-access-nqqnf\") pod \"auto-csr-approver-29556486-zprrc\" (UID: \"770aa8c1-c4df-4219-bad8-f77761708cb7\") " pod="openshift-infra/auto-csr-approver-29556486-zprrc" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.425823 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "643303f7-da91-41de-8fe6-407a7794de70" (UID: "643303f7-da91-41de-8fe6-407a7794de70"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.498063 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.498106 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643303f7-da91-41de-8fe6-407a7794de70-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.498120 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpcsm\" (UniqueName: \"kubernetes.io/projected/643303f7-da91-41de-8fe6-407a7794de70-kube-api-access-wpcsm\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:00 crc kubenswrapper[4876]: I0313 08:06:00.562533 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556486-zprrc" Mar 13 08:06:01 crc kubenswrapper[4876]: I0313 08:06:01.009734 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556486-zprrc"] Mar 13 08:06:01 crc kubenswrapper[4876]: W0313 08:06:01.017061 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod770aa8c1_c4df_4219_bad8_f77761708cb7.slice/crio-6ffa691b3707dffeede38f9a78c4785eb3f62ba1c81d1d7123e38c3865bd136d WatchSource:0}: Error finding container 6ffa691b3707dffeede38f9a78c4785eb3f62ba1c81d1d7123e38c3865bd136d: Status 404 returned error can't find the container with id 6ffa691b3707dffeede38f9a78c4785eb3f62ba1c81d1d7123e38c3865bd136d Mar 13 08:06:01 crc kubenswrapper[4876]: I0313 08:06:01.233800 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556486-zprrc" event={"ID":"770aa8c1-c4df-4219-bad8-f77761708cb7","Type":"ContainerStarted","Data":"6ffa691b3707dffeede38f9a78c4785eb3f62ba1c81d1d7123e38c3865bd136d"} Mar 13 08:06:01 crc kubenswrapper[4876]: I0313 08:06:01.233846 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9lsfg" Mar 13 08:06:01 crc kubenswrapper[4876]: I0313 08:06:01.276533 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lsfg"] Mar 13 08:06:01 crc kubenswrapper[4876]: I0313 08:06:01.293323 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9lsfg"] Mar 13 08:06:02 crc kubenswrapper[4876]: I0313 08:06:02.245319 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556486-zprrc" event={"ID":"770aa8c1-c4df-4219-bad8-f77761708cb7","Type":"ContainerStarted","Data":"31be08a512386fe4f1a39fa7ed417b42ad05e75f220adb51f58fb1cf7ae50103"} Mar 13 08:06:02 crc kubenswrapper[4876]: I0313 08:06:02.273708 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556486-zprrc" podStartSLOduration=1.499150267 podStartE2EDuration="2.273684578s" podCreationTimestamp="2026-03-13 08:06:00 +0000 UTC" firstStartedPulling="2026-03-13 08:06:01.018483011 +0000 UTC m=+1620.689261993" lastFinishedPulling="2026-03-13 08:06:01.793017322 +0000 UTC m=+1621.463796304" observedRunningTime="2026-03-13 08:06:02.263492868 +0000 UTC m=+1621.934271870" watchObservedRunningTime="2026-03-13 08:06:02.273684578 +0000 UTC m=+1621.944463560" Mar 13 08:06:03 crc kubenswrapper[4876]: I0313 08:06:03.048486 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="643303f7-da91-41de-8fe6-407a7794de70" path="/var/lib/kubelet/pods/643303f7-da91-41de-8fe6-407a7794de70/volumes" Mar 13 08:06:03 crc kubenswrapper[4876]: I0313 08:06:03.259348 4876 generic.go:334] "Generic (PLEG): container finished" podID="770aa8c1-c4df-4219-bad8-f77761708cb7" containerID="31be08a512386fe4f1a39fa7ed417b42ad05e75f220adb51f58fb1cf7ae50103" exitCode=0 Mar 13 08:06:03 crc kubenswrapper[4876]: I0313 08:06:03.259410 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556486-zprrc" event={"ID":"770aa8c1-c4df-4219-bad8-f77761708cb7","Type":"ContainerDied","Data":"31be08a512386fe4f1a39fa7ed417b42ad05e75f220adb51f58fb1cf7ae50103"} Mar 13 08:06:04 crc kubenswrapper[4876]: I0313 08:06:04.551082 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:06:04 crc kubenswrapper[4876]: I0313 08:06:04.552322 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:06:04 crc kubenswrapper[4876]: I0313 08:06:04.605589 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556486-zprrc" Mar 13 08:06:04 crc kubenswrapper[4876]: I0313 08:06:04.608606 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:06:04 crc kubenswrapper[4876]: I0313 08:06:04.787481 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqqnf\" (UniqueName: \"kubernetes.io/projected/770aa8c1-c4df-4219-bad8-f77761708cb7-kube-api-access-nqqnf\") pod \"770aa8c1-c4df-4219-bad8-f77761708cb7\" (UID: \"770aa8c1-c4df-4219-bad8-f77761708cb7\") " Mar 13 08:06:04 crc kubenswrapper[4876]: I0313 08:06:04.796049 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770aa8c1-c4df-4219-bad8-f77761708cb7-kube-api-access-nqqnf" (OuterVolumeSpecName: "kube-api-access-nqqnf") pod "770aa8c1-c4df-4219-bad8-f77761708cb7" (UID: "770aa8c1-c4df-4219-bad8-f77761708cb7"). InnerVolumeSpecName "kube-api-access-nqqnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:06:04 crc kubenswrapper[4876]: I0313 08:06:04.890423 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqqnf\" (UniqueName: \"kubernetes.io/projected/770aa8c1-c4df-4219-bad8-f77761708cb7-kube-api-access-nqqnf\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:05 crc kubenswrapper[4876]: I0313 08:06:05.278830 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556486-zprrc" Mar 13 08:06:05 crc kubenswrapper[4876]: I0313 08:06:05.278827 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556486-zprrc" event={"ID":"770aa8c1-c4df-4219-bad8-f77761708cb7","Type":"ContainerDied","Data":"6ffa691b3707dffeede38f9a78c4785eb3f62ba1c81d1d7123e38c3865bd136d"} Mar 13 08:06:05 crc kubenswrapper[4876]: I0313 08:06:05.278894 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ffa691b3707dffeede38f9a78c4785eb3f62ba1c81d1d7123e38c3865bd136d" Mar 13 08:06:05 crc kubenswrapper[4876]: I0313 08:06:05.343980 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556480-8292c"] Mar 13 08:06:05 crc kubenswrapper[4876]: I0313 08:06:05.346024 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:06:05 crc kubenswrapper[4876]: I0313 08:06:05.369164 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556480-8292c"] Mar 13 08:06:05 crc kubenswrapper[4876]: I0313 08:06:05.784202 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqccs"] Mar 13 08:06:06 crc kubenswrapper[4876]: E0313 08:06:06.359677 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod643303f7_da91_41de_8fe6_407a7794de70.slice\": RecentStats: unable to find data in memory cache]" Mar 13 08:06:07 crc kubenswrapper[4876]: I0313 08:06:07.048705 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="306b6d6a-7ec0-41f2-b7e6-8974bb873700" path="/var/lib/kubelet/pods/306b6d6a-7ec0-41f2-b7e6-8974bb873700/volumes" Mar 13 08:06:07 crc kubenswrapper[4876]: I0313 08:06:07.298744 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zqccs" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerName="registry-server" containerID="cri-o://3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0" gracePeriod=2 Mar 13 08:06:07 crc kubenswrapper[4876]: I0313 08:06:07.860790 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:06:07 crc kubenswrapper[4876]: I0313 08:06:07.961157 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv5z8\" (UniqueName: \"kubernetes.io/projected/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-kube-api-access-vv5z8\") pod \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " Mar 13 08:06:07 crc kubenswrapper[4876]: I0313 08:06:07.969515 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-kube-api-access-vv5z8" (OuterVolumeSpecName: "kube-api-access-vv5z8") pod "0a123dd8-46d2-4bf9-92a7-34151f54c0e1" (UID: "0a123dd8-46d2-4bf9-92a7-34151f54c0e1"). InnerVolumeSpecName "kube-api-access-vv5z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.035709 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:06:08 crc kubenswrapper[4876]: E0313 08:06:08.036164 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.063475 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-utilities\") pod \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.063677 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-catalog-content\") pod \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\" (UID: \"0a123dd8-46d2-4bf9-92a7-34151f54c0e1\") " Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.064284 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-utilities" (OuterVolumeSpecName: "utilities") pod "0a123dd8-46d2-4bf9-92a7-34151f54c0e1" (UID: "0a123dd8-46d2-4bf9-92a7-34151f54c0e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.064546 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv5z8\" (UniqueName: \"kubernetes.io/projected/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-kube-api-access-vv5z8\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.064577 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.316949 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerID="3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0" exitCode=0 Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.317034 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqccs" event={"ID":"0a123dd8-46d2-4bf9-92a7-34151f54c0e1","Type":"ContainerDied","Data":"3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0"} Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.317086 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqccs" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.317099 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqccs" event={"ID":"0a123dd8-46d2-4bf9-92a7-34151f54c0e1","Type":"ContainerDied","Data":"40ce537937db57e137ac20976f94440080877e972ddf0eab0c158b25188196a0"} Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.317124 4876 scope.go:117] "RemoveContainer" containerID="3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.346088 4876 scope.go:117] "RemoveContainer" containerID="3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.377520 4876 scope.go:117] "RemoveContainer" containerID="58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.423388 4876 scope.go:117] "RemoveContainer" containerID="3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0" Mar 13 08:06:08 crc kubenswrapper[4876]: E0313 08:06:08.424864 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0\": container with ID starting with 3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0 not found: ID does not exist" containerID="3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.424948 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0"} err="failed to get container status \"3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0\": rpc error: code = NotFound desc = could not find container \"3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0\": container with ID starting with 3f33788dbf5e438684e95bf564668a3b4799a48a9e3f5a4ea7addf863e2f64c0 not found: ID does not exist" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.424988 4876 scope.go:117] "RemoveContainer" containerID="3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809" Mar 13 08:06:08 crc kubenswrapper[4876]: E0313 08:06:08.425435 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809\": container with ID starting with 3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809 not found: ID does not exist" containerID="3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.425465 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809"} err="failed to get container status \"3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809\": rpc error: code = NotFound desc = could not find container \"3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809\": container with ID starting with 3c08fc0a34f1a203befa2f5fa6d2e4c2de05b115aad3bfcad7c04099e89d0809 not found: ID does not exist" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.425480 4876 scope.go:117] "RemoveContainer" containerID="58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda" Mar 13 08:06:08 crc kubenswrapper[4876]: E0313 08:06:08.426164 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda\": container with ID starting with 58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda not found: ID does not exist" containerID="58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.426203 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda"} err="failed to get container status \"58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda\": rpc error: code = NotFound desc = could not find container \"58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda\": container with ID starting with 58b18800357423bbab179a5f9075eca4a9aec73ca0cd01f91a3fb105849cbcda not found: ID does not exist" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.457291 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a123dd8-46d2-4bf9-92a7-34151f54c0e1" (UID: "0a123dd8-46d2-4bf9-92a7-34151f54c0e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.477273 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a123dd8-46d2-4bf9-92a7-34151f54c0e1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.654388 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqccs"] Mar 13 08:06:08 crc kubenswrapper[4876]: I0313 08:06:08.667420 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zqccs"] Mar 13 08:06:09 crc kubenswrapper[4876]: I0313 08:06:09.046839 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" path="/var/lib/kubelet/pods/0a123dd8-46d2-4bf9-92a7-34151f54c0e1/volumes" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.212068 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4nfq7"] Mar 13 08:06:11 crc kubenswrapper[4876]: E0313 08:06:11.212947 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643303f7-da91-41de-8fe6-407a7794de70" containerName="extract-utilities" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.212964 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="643303f7-da91-41de-8fe6-407a7794de70" containerName="extract-utilities" Mar 13 08:06:11 crc kubenswrapper[4876]: E0313 08:06:11.212981 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770aa8c1-c4df-4219-bad8-f77761708cb7" containerName="oc" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.212989 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="770aa8c1-c4df-4219-bad8-f77761708cb7" containerName="oc" Mar 13 08:06:11 crc kubenswrapper[4876]: E0313 08:06:11.213005 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerName="registry-server" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.213015 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerName="registry-server" Mar 13 08:06:11 crc kubenswrapper[4876]: E0313 08:06:11.213048 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643303f7-da91-41de-8fe6-407a7794de70" containerName="extract-content" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.213055 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="643303f7-da91-41de-8fe6-407a7794de70" containerName="extract-content" Mar 13 08:06:11 crc kubenswrapper[4876]: E0313 08:06:11.213068 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerName="extract-utilities" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.213075 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerName="extract-utilities" Mar 13 08:06:11 crc kubenswrapper[4876]: E0313 08:06:11.213093 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643303f7-da91-41de-8fe6-407a7794de70" containerName="registry-server" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.213099 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="643303f7-da91-41de-8fe6-407a7794de70" containerName="registry-server" Mar 13 08:06:11 crc kubenswrapper[4876]: E0313 08:06:11.213119 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerName="extract-content" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.213125 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerName="extract-content" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.213406 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="643303f7-da91-41de-8fe6-407a7794de70" containerName="registry-server" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.213420 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="770aa8c1-c4df-4219-bad8-f77761708cb7" containerName="oc" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.213437 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a123dd8-46d2-4bf9-92a7-34151f54c0e1" containerName="registry-server" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.215171 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.224253 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4nfq7"] Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.338088 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-catalog-content\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.338334 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-utilities\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.338418 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnf5f\" (UniqueName: \"kubernetes.io/projected/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-kube-api-access-gnf5f\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.440819 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-utilities\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.440914 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnf5f\" (UniqueName: \"kubernetes.io/projected/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-kube-api-access-gnf5f\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.441003 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-catalog-content\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.441761 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-catalog-content\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.442047 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-utilities\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.464276 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnf5f\" (UniqueName: \"kubernetes.io/projected/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-kube-api-access-gnf5f\") pod \"certified-operators-4nfq7\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:11 crc kubenswrapper[4876]: I0313 08:06:11.543465 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:12 crc kubenswrapper[4876]: I0313 08:06:12.070849 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4nfq7"] Mar 13 08:06:12 crc kubenswrapper[4876]: W0313 08:06:12.076468 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod030bec3c_f2dc_4c7a_96a9_93c90dd92db4.slice/crio-971bd6daf3897dc996d1736ef1d9ab857371caef478433caa82eec28529c1a65 WatchSource:0}: Error finding container 971bd6daf3897dc996d1736ef1d9ab857371caef478433caa82eec28529c1a65: Status 404 returned error can't find the container with id 971bd6daf3897dc996d1736ef1d9ab857371caef478433caa82eec28529c1a65 Mar 13 08:06:12 crc kubenswrapper[4876]: I0313 08:06:12.362260 4876 generic.go:334] "Generic (PLEG): container finished" podID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerID="e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31" exitCode=0 Mar 13 08:06:12 crc kubenswrapper[4876]: I0313 08:06:12.362373 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nfq7" event={"ID":"030bec3c-f2dc-4c7a-96a9-93c90dd92db4","Type":"ContainerDied","Data":"e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31"} Mar 13 08:06:12 crc kubenswrapper[4876]: I0313 08:06:12.362637 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nfq7" event={"ID":"030bec3c-f2dc-4c7a-96a9-93c90dd92db4","Type":"ContainerStarted","Data":"971bd6daf3897dc996d1736ef1d9ab857371caef478433caa82eec28529c1a65"} Mar 13 08:06:13 crc kubenswrapper[4876]: I0313 08:06:13.376397 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nfq7" event={"ID":"030bec3c-f2dc-4c7a-96a9-93c90dd92db4","Type":"ContainerStarted","Data":"577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e"} Mar 13 08:06:14 crc kubenswrapper[4876]: I0313 08:06:14.389413 4876 generic.go:334] "Generic (PLEG): container finished" podID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerID="577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e" exitCode=0 Mar 13 08:06:14 crc kubenswrapper[4876]: I0313 08:06:14.389464 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nfq7" event={"ID":"030bec3c-f2dc-4c7a-96a9-93c90dd92db4","Type":"ContainerDied","Data":"577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e"} Mar 13 08:06:15 crc kubenswrapper[4876]: I0313 08:06:15.400632 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nfq7" event={"ID":"030bec3c-f2dc-4c7a-96a9-93c90dd92db4","Type":"ContainerStarted","Data":"b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44"} Mar 13 08:06:15 crc kubenswrapper[4876]: I0313 08:06:15.425818 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4nfq7" podStartSLOduration=1.757726782 podStartE2EDuration="4.425792179s" podCreationTimestamp="2026-03-13 08:06:11 +0000 UTC" firstStartedPulling="2026-03-13 08:06:12.363844866 +0000 UTC m=+1632.034623848" lastFinishedPulling="2026-03-13 08:06:15.031910253 +0000 UTC m=+1634.702689245" observedRunningTime="2026-03-13 08:06:15.420662418 +0000 UTC m=+1635.091441400" watchObservedRunningTime="2026-03-13 08:06:15.425792179 +0000 UTC m=+1635.096571161" Mar 13 08:06:16 crc kubenswrapper[4876]: E0313 08:06:16.643765 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod643303f7_da91_41de_8fe6_407a7794de70.slice\": RecentStats: unable to find data in memory cache]" Mar 13 08:06:19 crc kubenswrapper[4876]: I0313 08:06:19.725069 4876 scope.go:117] "RemoveContainer" containerID="8894aa4f85279d09acd15fc0d197c863bf3135ff03572334180dd504ffd1e85d" Mar 13 08:06:19 crc kubenswrapper[4876]: I0313 08:06:19.767774 4876 scope.go:117] "RemoveContainer" containerID="c470d21baa2b3e7224c1b2bdee256413e3599b9b1b94c809bd7d9a21e8a21dfc" Mar 13 08:06:19 crc kubenswrapper[4876]: I0313 08:06:19.814326 4876 scope.go:117] "RemoveContainer" containerID="4555ee39a98e78d790cba89dd0f0321a88a7e01669427190e0b33f208846da3e" Mar 13 08:06:19 crc kubenswrapper[4876]: I0313 08:06:19.853486 4876 scope.go:117] "RemoveContainer" containerID="8065d8f3b387985421911bcc02b9ff1f0f13eca5ea11e59670b06fe74e46e7cd" Mar 13 08:06:19 crc kubenswrapper[4876]: I0313 08:06:19.879899 4876 scope.go:117] "RemoveContainer" containerID="bba9f685500d5103dcfef11e9fbf6d51e86619e17f8245e274c3009e4a681e8a" Mar 13 08:06:21 crc kubenswrapper[4876]: I0313 08:06:21.544146 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:21 crc kubenswrapper[4876]: I0313 08:06:21.544605 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:21 crc kubenswrapper[4876]: I0313 08:06:21.594909 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:22 crc kubenswrapper[4876]: I0313 08:06:22.537992 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:22 crc kubenswrapper[4876]: I0313 08:06:22.592500 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4nfq7"] Mar 13 08:06:23 crc kubenswrapper[4876]: I0313 08:06:23.037851 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:06:23 crc kubenswrapper[4876]: E0313 08:06:23.040076 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:06:24 crc kubenswrapper[4876]: I0313 08:06:24.497702 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4nfq7" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerName="registry-server" containerID="cri-o://b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44" gracePeriod=2 Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.003347 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.128931 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-utilities\") pod \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.129279 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-catalog-content\") pod \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.129367 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnf5f\" (UniqueName: \"kubernetes.io/projected/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-kube-api-access-gnf5f\") pod \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\" (UID: \"030bec3c-f2dc-4c7a-96a9-93c90dd92db4\") " Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.129954 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-utilities" (OuterVolumeSpecName: "utilities") pod "030bec3c-f2dc-4c7a-96a9-93c90dd92db4" (UID: "030bec3c-f2dc-4c7a-96a9-93c90dd92db4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.134629 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.136948 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-kube-api-access-gnf5f" (OuterVolumeSpecName: "kube-api-access-gnf5f") pod "030bec3c-f2dc-4c7a-96a9-93c90dd92db4" (UID: "030bec3c-f2dc-4c7a-96a9-93c90dd92db4"). InnerVolumeSpecName "kube-api-access-gnf5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.192934 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "030bec3c-f2dc-4c7a-96a9-93c90dd92db4" (UID: "030bec3c-f2dc-4c7a-96a9-93c90dd92db4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.237144 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.237204 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnf5f\" (UniqueName: \"kubernetes.io/projected/030bec3c-f2dc-4c7a-96a9-93c90dd92db4-kube-api-access-gnf5f\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.532506 4876 generic.go:334] "Generic (PLEG): container finished" podID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerID="b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44" exitCode=0 Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.532593 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4nfq7" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.532585 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nfq7" event={"ID":"030bec3c-f2dc-4c7a-96a9-93c90dd92db4","Type":"ContainerDied","Data":"b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44"} Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.532681 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4nfq7" event={"ID":"030bec3c-f2dc-4c7a-96a9-93c90dd92db4","Type":"ContainerDied","Data":"971bd6daf3897dc996d1736ef1d9ab857371caef478433caa82eec28529c1a65"} Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.532711 4876 scope.go:117] "RemoveContainer" containerID="b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.564475 4876 scope.go:117] "RemoveContainer" containerID="577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.575635 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4nfq7"] Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.586661 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4nfq7"] Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.591577 4876 scope.go:117] "RemoveContainer" containerID="e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.651497 4876 scope.go:117] "RemoveContainer" containerID="b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44" Mar 13 08:06:25 crc kubenswrapper[4876]: E0313 08:06:25.651956 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44\": container with ID starting with b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44 not found: ID does not exist" containerID="b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.652036 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44"} err="failed to get container status \"b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44\": rpc error: code = NotFound desc = could not find container \"b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44\": container with ID starting with b7efbcee4df0a666e46ac862422424214ca7f0cec005c88dd82fe4e094c05d44 not found: ID does not exist" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.652065 4876 scope.go:117] "RemoveContainer" containerID="577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e" Mar 13 08:06:25 crc kubenswrapper[4876]: E0313 08:06:25.652536 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e\": container with ID starting with 577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e not found: ID does not exist" containerID="577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.652560 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e"} err="failed to get container status \"577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e\": rpc error: code = NotFound desc = could not find container \"577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e\": container with ID starting with 577de7065401324f0d6c3d7d465796baa0fc6344ddef14c3ef72738274f0606e not found: ID does not exist" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.652576 4876 scope.go:117] "RemoveContainer" containerID="e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31" Mar 13 08:06:25 crc kubenswrapper[4876]: E0313 08:06:25.653175 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31\": container with ID starting with e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31 not found: ID does not exist" containerID="e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31" Mar 13 08:06:25 crc kubenswrapper[4876]: I0313 08:06:25.653202 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31"} err="failed to get container status \"e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31\": rpc error: code = NotFound desc = could not find container \"e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31\": container with ID starting with e9b44b4473237986b5a22c16b13ca2fd5926676bde1420ff31ab42a7fa6add31 not found: ID does not exist" Mar 13 08:06:26 crc kubenswrapper[4876]: E0313 08:06:26.913640 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod643303f7_da91_41de_8fe6_407a7794de70.slice\": RecentStats: unable to find data in memory cache]" Mar 13 08:06:27 crc kubenswrapper[4876]: I0313 08:06:27.048910 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" path="/var/lib/kubelet/pods/030bec3c-f2dc-4c7a-96a9-93c90dd92db4/volumes" Mar 13 08:06:35 crc kubenswrapper[4876]: I0313 08:06:35.036155 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:06:35 crc kubenswrapper[4876]: E0313 08:06:35.037528 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:06:37 crc kubenswrapper[4876]: E0313 08:06:37.183041 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod643303f7_da91_41de_8fe6_407a7794de70.slice\": RecentStats: unable to find data in memory cache]" Mar 13 08:06:47 crc kubenswrapper[4876]: E0313 08:06:47.455103 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod643303f7_da91_41de_8fe6_407a7794de70.slice\": RecentStats: unable to find data in memory cache]" Mar 13 08:06:48 crc kubenswrapper[4876]: I0313 08:06:48.036015 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:06:48 crc kubenswrapper[4876]: E0313 08:06:48.036573 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:06:56 crc kubenswrapper[4876]: I0313 08:06:56.920667 4876 generic.go:334] "Generic (PLEG): container finished" podID="0808c5a4-53f5-43e5-af39-bf78530e1d71" containerID="474f8ca4ddf1c9df45575afff5bb96e4d8f7a63f3f522f2d40842f4ac083fe97" exitCode=0 Mar 13 08:06:56 crc kubenswrapper[4876]: I0313 08:06:56.920738 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" event={"ID":"0808c5a4-53f5-43e5-af39-bf78530e1d71","Type":"ContainerDied","Data":"474f8ca4ddf1c9df45575afff5bb96e4d8f7a63f3f522f2d40842f4ac083fe97"} Mar 13 08:06:57 crc kubenswrapper[4876]: E0313 08:06:57.694632 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod643303f7_da91_41de_8fe6_407a7794de70.slice\": RecentStats: unable to find data in memory cache]" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.367354 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.455308 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-bootstrap-combined-ca-bundle\") pod \"0808c5a4-53f5-43e5-af39-bf78530e1d71\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.455385 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-inventory\") pod \"0808c5a4-53f5-43e5-af39-bf78530e1d71\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.455426 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-ssh-key-openstack-edpm-ipam\") pod \"0808c5a4-53f5-43e5-af39-bf78530e1d71\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.455532 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt6cq\" (UniqueName: \"kubernetes.io/projected/0808c5a4-53f5-43e5-af39-bf78530e1d71-kube-api-access-bt6cq\") pod \"0808c5a4-53f5-43e5-af39-bf78530e1d71\" (UID: \"0808c5a4-53f5-43e5-af39-bf78530e1d71\") " Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.461703 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0808c5a4-53f5-43e5-af39-bf78530e1d71" (UID: "0808c5a4-53f5-43e5-af39-bf78530e1d71"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.462382 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0808c5a4-53f5-43e5-af39-bf78530e1d71-kube-api-access-bt6cq" (OuterVolumeSpecName: "kube-api-access-bt6cq") pod "0808c5a4-53f5-43e5-af39-bf78530e1d71" (UID: "0808c5a4-53f5-43e5-af39-bf78530e1d71"). InnerVolumeSpecName "kube-api-access-bt6cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.490578 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0808c5a4-53f5-43e5-af39-bf78530e1d71" (UID: "0808c5a4-53f5-43e5-af39-bf78530e1d71"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.490790 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-inventory" (OuterVolumeSpecName: "inventory") pod "0808c5a4-53f5-43e5-af39-bf78530e1d71" (UID: "0808c5a4-53f5-43e5-af39-bf78530e1d71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.558732 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt6cq\" (UniqueName: \"kubernetes.io/projected/0808c5a4-53f5-43e5-af39-bf78530e1d71-kube-api-access-bt6cq\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.558784 4876 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.558796 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.558805 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0808c5a4-53f5-43e5-af39-bf78530e1d71-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.945645 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" event={"ID":"0808c5a4-53f5-43e5-af39-bf78530e1d71","Type":"ContainerDied","Data":"cba143fd90d7931f15ad699d1f38a8462866f887d1069d5aae7d6990ad8e7c4b"} Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.945696 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cba143fd90d7931f15ad699d1f38a8462866f887d1069d5aae7d6990ad8e7c4b" Mar 13 08:06:58 crc kubenswrapper[4876]: I0313 08:06:58.945758 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.052055 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht"] Mar 13 08:06:59 crc kubenswrapper[4876]: E0313 08:06:59.052772 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerName="registry-server" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.052906 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerName="registry-server" Mar 13 08:06:59 crc kubenswrapper[4876]: E0313 08:06:59.053024 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0808c5a4-53f5-43e5-af39-bf78530e1d71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.053136 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0808c5a4-53f5-43e5-af39-bf78530e1d71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 08:06:59 crc kubenswrapper[4876]: E0313 08:06:59.053338 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerName="extract-content" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.053460 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerName="extract-content" Mar 13 08:06:59 crc kubenswrapper[4876]: E0313 08:06:59.053548 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerName="extract-utilities" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.053644 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerName="extract-utilities" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.053998 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0808c5a4-53f5-43e5-af39-bf78530e1d71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.054107 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="030bec3c-f2dc-4c7a-96a9-93c90dd92db4" containerName="registry-server" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.055094 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.057687 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.058338 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.058476 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.058349 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.071830 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht"] Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.171826 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.172217 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.172343 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djtmj\" (UniqueName: \"kubernetes.io/projected/0de3f5bc-8761-491e-9a81-9a03e796ca1b-kube-api-access-djtmj\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.274050 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.274155 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.274218 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djtmj\" (UniqueName: \"kubernetes.io/projected/0de3f5bc-8761-491e-9a81-9a03e796ca1b-kube-api-access-djtmj\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.279977 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.280809 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.294923 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djtmj\" (UniqueName: \"kubernetes.io/projected/0de3f5bc-8761-491e-9a81-9a03e796ca1b-kube-api-access-djtmj\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fwnht\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.374830 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:06:59 crc kubenswrapper[4876]: I0313 08:06:59.988946 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht"] Mar 13 08:07:00 crc kubenswrapper[4876]: I0313 08:07:00.970304 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" event={"ID":"0de3f5bc-8761-491e-9a81-9a03e796ca1b","Type":"ContainerStarted","Data":"eec6714d8573d91b942761bdbc29b6f14f03c598c5f7cd0001cfe1d2b9443e50"} Mar 13 08:07:00 crc kubenswrapper[4876]: I0313 08:07:00.970915 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" event={"ID":"0de3f5bc-8761-491e-9a81-9a03e796ca1b","Type":"ContainerStarted","Data":"f68016a53c1b011a78572fbd55948351591243e2cd97a0b5cebacfd2a4d955f0"} Mar 13 08:07:00 crc kubenswrapper[4876]: I0313 08:07:00.994186 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" podStartSLOduration=1.513373866 podStartE2EDuration="1.994155855s" podCreationTimestamp="2026-03-13 08:06:59 +0000 UTC" firstStartedPulling="2026-03-13 08:07:00.002087591 +0000 UTC m=+1679.672866583" lastFinishedPulling="2026-03-13 08:07:00.48286959 +0000 UTC m=+1680.153648572" observedRunningTime="2026-03-13 08:07:00.988753058 +0000 UTC m=+1680.659532040" watchObservedRunningTime="2026-03-13 08:07:00.994155855 +0000 UTC m=+1680.664934857" Mar 13 08:07:03 crc kubenswrapper[4876]: I0313 08:07:03.036853 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:07:03 crc kubenswrapper[4876]: E0313 08:07:03.037553 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:07:18 crc kubenswrapper[4876]: I0313 08:07:18.035177 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:07:18 crc kubenswrapper[4876]: E0313 08:07:18.036020 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:07:19 crc kubenswrapper[4876]: I0313 08:07:19.981076 4876 scope.go:117] "RemoveContainer" containerID="78f4e644173b7f3302894f0ad9053e390bf7f744062e95401dc0f8243906be39" Mar 13 08:07:20 crc kubenswrapper[4876]: I0313 08:07:20.021167 4876 scope.go:117] "RemoveContainer" containerID="dcca7debd5b8cd22cc56204986eb23b1a779774e0c8759ef5f99a62a405aa6b8" Mar 13 08:07:20 crc kubenswrapper[4876]: I0313 08:07:20.047652 4876 scope.go:117] "RemoveContainer" containerID="cb1c39bcd6bcd4f303c36078f194ac7f5958f4b970935b4a99c5dc3d2fc3c662" Mar 13 08:07:33 crc kubenswrapper[4876]: I0313 08:07:33.037061 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:07:33 crc kubenswrapper[4876]: E0313 08:07:33.038820 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:07:46 crc kubenswrapper[4876]: I0313 08:07:46.036367 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:07:46 crc kubenswrapper[4876]: E0313 08:07:46.037647 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:07:46 crc kubenswrapper[4876]: I0313 08:07:46.050692 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-t8c2h"] Mar 13 08:07:46 crc kubenswrapper[4876]: I0313 08:07:46.062317 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-70c0-account-create-update-bvqs5"] Mar 13 08:07:46 crc kubenswrapper[4876]: I0313 08:07:46.072083 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-t8c2h"] Mar 13 08:07:46 crc kubenswrapper[4876]: I0313 08:07:46.081661 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-70c0-account-create-update-bvqs5"] Mar 13 08:07:47 crc kubenswrapper[4876]: I0313 08:07:47.046888 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0341c6d0-9f56-4dcd-a167-7152687cf2c1" path="/var/lib/kubelet/pods/0341c6d0-9f56-4dcd-a167-7152687cf2c1/volumes" Mar 13 08:07:47 crc kubenswrapper[4876]: I0313 08:07:47.047751 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61f27a21-4172-4552-a738-314cceb0acd7" path="/var/lib/kubelet/pods/61f27a21-4172-4552-a738-314cceb0acd7/volumes" Mar 13 08:07:52 crc kubenswrapper[4876]: I0313 08:07:52.038925 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-bksxd"] Mar 13 08:07:52 crc kubenswrapper[4876]: I0313 08:07:52.049273 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-d7a7-account-create-update-q8h2f"] Mar 13 08:07:52 crc kubenswrapper[4876]: I0313 08:07:52.057877 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-d7a7-account-create-update-q8h2f"] Mar 13 08:07:52 crc kubenswrapper[4876]: I0313 08:07:52.065529 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-bksxd"] Mar 13 08:07:53 crc kubenswrapper[4876]: I0313 08:07:53.056108 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="891ee27b-6971-41e8-baee-ea2f4b4a0a4e" path="/var/lib/kubelet/pods/891ee27b-6971-41e8-baee-ea2f4b4a0a4e/volumes" Mar 13 08:07:53 crc kubenswrapper[4876]: I0313 08:07:53.059633 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4188868-e4dc-4960-a300-ebc151e4408f" path="/var/lib/kubelet/pods/f4188868-e4dc-4960-a300-ebc151e4408f/volumes" Mar 13 08:07:54 crc kubenswrapper[4876]: I0313 08:07:54.033903 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jkf72"] Mar 13 08:07:54 crc kubenswrapper[4876]: I0313 08:07:54.047797 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8359-account-create-update-fr5tj"] Mar 13 08:07:54 crc kubenswrapper[4876]: I0313 08:07:54.059706 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8359-account-create-update-fr5tj"] Mar 13 08:07:54 crc kubenswrapper[4876]: I0313 08:07:54.069450 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jkf72"] Mar 13 08:07:55 crc kubenswrapper[4876]: I0313 08:07:55.047338 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13197959-749d-4222-ae6c-36fe74c289ff" path="/var/lib/kubelet/pods/13197959-749d-4222-ae6c-36fe74c289ff/volumes" Mar 13 08:07:55 crc kubenswrapper[4876]: I0313 08:07:55.048023 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32838d97-317e-4102-9fdc-d707329a4f45" path="/var/lib/kubelet/pods/32838d97-317e-4102-9fdc-d707329a4f45/volumes" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.036832 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:08:00 crc kubenswrapper[4876]: E0313 08:08:00.037890 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.156564 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556488-mwrfn"] Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.158477 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556488-mwrfn" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.163791 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.165898 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.165930 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.166462 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556488-mwrfn"] Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.323948 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxmsg\" (UniqueName: \"kubernetes.io/projected/5452b7b8-0dbd-46f7-91d4-2321bcb62f5f-kube-api-access-nxmsg\") pod \"auto-csr-approver-29556488-mwrfn\" (UID: \"5452b7b8-0dbd-46f7-91d4-2321bcb62f5f\") " pod="openshift-infra/auto-csr-approver-29556488-mwrfn" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.425740 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxmsg\" (UniqueName: \"kubernetes.io/projected/5452b7b8-0dbd-46f7-91d4-2321bcb62f5f-kube-api-access-nxmsg\") pod \"auto-csr-approver-29556488-mwrfn\" (UID: \"5452b7b8-0dbd-46f7-91d4-2321bcb62f5f\") " pod="openshift-infra/auto-csr-approver-29556488-mwrfn" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.445278 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxmsg\" (UniqueName: \"kubernetes.io/projected/5452b7b8-0dbd-46f7-91d4-2321bcb62f5f-kube-api-access-nxmsg\") pod \"auto-csr-approver-29556488-mwrfn\" (UID: \"5452b7b8-0dbd-46f7-91d4-2321bcb62f5f\") " pod="openshift-infra/auto-csr-approver-29556488-mwrfn" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.484898 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556488-mwrfn" Mar 13 08:08:00 crc kubenswrapper[4876]: I0313 08:08:00.982160 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556488-mwrfn"] Mar 13 08:08:01 crc kubenswrapper[4876]: I0313 08:08:01.691948 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556488-mwrfn" event={"ID":"5452b7b8-0dbd-46f7-91d4-2321bcb62f5f","Type":"ContainerStarted","Data":"e115a00d827f192ee2704b085b064d8e7726a9537d8c304b0f58686f8a718d50"} Mar 13 08:08:02 crc kubenswrapper[4876]: I0313 08:08:02.705962 4876 generic.go:334] "Generic (PLEG): container finished" podID="5452b7b8-0dbd-46f7-91d4-2321bcb62f5f" containerID="8c2d6cd129b5cf087e15db2b85a5464636c60bfb51c76696d6f692964d49e6aa" exitCode=0 Mar 13 08:08:02 crc kubenswrapper[4876]: I0313 08:08:02.706067 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556488-mwrfn" event={"ID":"5452b7b8-0dbd-46f7-91d4-2321bcb62f5f","Type":"ContainerDied","Data":"8c2d6cd129b5cf087e15db2b85a5464636c60bfb51c76696d6f692964d49e6aa"} Mar 13 08:08:04 crc kubenswrapper[4876]: I0313 08:08:04.030153 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556488-mwrfn" Mar 13 08:08:04 crc kubenswrapper[4876]: I0313 08:08:04.212101 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxmsg\" (UniqueName: \"kubernetes.io/projected/5452b7b8-0dbd-46f7-91d4-2321bcb62f5f-kube-api-access-nxmsg\") pod \"5452b7b8-0dbd-46f7-91d4-2321bcb62f5f\" (UID: \"5452b7b8-0dbd-46f7-91d4-2321bcb62f5f\") " Mar 13 08:08:04 crc kubenswrapper[4876]: I0313 08:08:04.220741 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5452b7b8-0dbd-46f7-91d4-2321bcb62f5f-kube-api-access-nxmsg" (OuterVolumeSpecName: "kube-api-access-nxmsg") pod "5452b7b8-0dbd-46f7-91d4-2321bcb62f5f" (UID: "5452b7b8-0dbd-46f7-91d4-2321bcb62f5f"). InnerVolumeSpecName "kube-api-access-nxmsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:08:04 crc kubenswrapper[4876]: I0313 08:08:04.314573 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxmsg\" (UniqueName: \"kubernetes.io/projected/5452b7b8-0dbd-46f7-91d4-2321bcb62f5f-kube-api-access-nxmsg\") on node \"crc\" DevicePath \"\"" Mar 13 08:08:04 crc kubenswrapper[4876]: I0313 08:08:04.725858 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556488-mwrfn" event={"ID":"5452b7b8-0dbd-46f7-91d4-2321bcb62f5f","Type":"ContainerDied","Data":"e115a00d827f192ee2704b085b064d8e7726a9537d8c304b0f58686f8a718d50"} Mar 13 08:08:04 crc kubenswrapper[4876]: I0313 08:08:04.725903 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556488-mwrfn" Mar 13 08:08:04 crc kubenswrapper[4876]: I0313 08:08:04.725908 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e115a00d827f192ee2704b085b064d8e7726a9537d8c304b0f58686f8a718d50" Mar 13 08:08:05 crc kubenswrapper[4876]: I0313 08:08:05.102901 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556482-85shb"] Mar 13 08:08:05 crc kubenswrapper[4876]: I0313 08:08:05.113179 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556482-85shb"] Mar 13 08:08:07 crc kubenswrapper[4876]: I0313 08:08:07.056330 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d26e6dde-879c-4a43-98a4-cf908aa696f7" path="/var/lib/kubelet/pods/d26e6dde-879c-4a43-98a4-cf908aa696f7/volumes" Mar 13 08:08:12 crc kubenswrapper[4876]: I0313 08:08:12.030130 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-kc84d"] Mar 13 08:08:12 crc kubenswrapper[4876]: I0313 08:08:12.036223 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:08:12 crc kubenswrapper[4876]: E0313 08:08:12.036668 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:08:12 crc kubenswrapper[4876]: I0313 08:08:12.038195 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-kc84d"] Mar 13 08:08:13 crc kubenswrapper[4876]: I0313 08:08:13.053227 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7213beb-842a-408a-8555-57d912b99703" path="/var/lib/kubelet/pods/c7213beb-842a-408a-8555-57d912b99703/volumes" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.057740 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-ccnjk"] Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.069744 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-ccnjk"] Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.129272 4876 scope.go:117] "RemoveContainer" containerID="c4a646bbc62a1521689dd3bef40edaa23378fa7c0632719d93ff39de8669a54c" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.175374 4876 scope.go:117] "RemoveContainer" containerID="f922a95faaaf45391c12a6d61c2a2671af6fa3251036f30d410a0d7a480ad2e4" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.217271 4876 scope.go:117] "RemoveContainer" containerID="81f6aba628fb9a241fd41b029fbc0358f3eb5a533fbdcebba9a35a9c734df151" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.261130 4876 scope.go:117] "RemoveContainer" containerID="22cf8295b2ad80a74b8ac07c17b19ea2e5caeb2f8fde49408278ee1dd60f0295" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.289114 4876 scope.go:117] "RemoveContainer" containerID="ab93520b745adca95f5146d9af5af7c14cf6266c8ba8465da18f8188ec64d2b0" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.363820 4876 scope.go:117] "RemoveContainer" containerID="b950aa6b5a6a0e476622df730068e6056cc98e7ae5ec28122afd0b1ee7cc581e" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.393408 4876 scope.go:117] "RemoveContainer" containerID="a8f830fcaa939cb42366cb37d1e9ebaed6171d66f16a8e1d4ab81fa1711c3eda" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.451447 4876 scope.go:117] "RemoveContainer" containerID="4b791e896a8d771005979cef57e910c008e7ac361f29800c3e151a8fbeca0751" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.481319 4876 scope.go:117] "RemoveContainer" containerID="79fd805194e9fd15bbe81ceb24d961ed287f1fd670b5c8fa45a2bdb63466873e" Mar 13 08:08:20 crc kubenswrapper[4876]: I0313 08:08:20.509381 4876 scope.go:117] "RemoveContainer" containerID="4f9013fb4dc8dee9f4e68587e3f4df7c9ea1d9d5752bd42e435e22f98ff21602" Mar 13 08:08:21 crc kubenswrapper[4876]: I0313 08:08:21.049995 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="275d5189-c7c8-44f4-b753-9590b692900c" path="/var/lib/kubelet/pods/275d5189-c7c8-44f4-b753-9590b692900c/volumes" Mar 13 08:08:24 crc kubenswrapper[4876]: I0313 08:08:24.929229 4876 generic.go:334] "Generic (PLEG): container finished" podID="0de3f5bc-8761-491e-9a81-9a03e796ca1b" containerID="eec6714d8573d91b942761bdbc29b6f14f03c598c5f7cd0001cfe1d2b9443e50" exitCode=0 Mar 13 08:08:24 crc kubenswrapper[4876]: I0313 08:08:24.929296 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" event={"ID":"0de3f5bc-8761-491e-9a81-9a03e796ca1b","Type":"ContainerDied","Data":"eec6714d8573d91b942761bdbc29b6f14f03c598c5f7cd0001cfe1d2b9443e50"} Mar 13 08:08:25 crc kubenswrapper[4876]: I0313 08:08:25.036373 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:08:25 crc kubenswrapper[4876]: E0313 08:08:25.036613 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.438953 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.542959 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-ssh-key-openstack-edpm-ipam\") pod \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.543421 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djtmj\" (UniqueName: \"kubernetes.io/projected/0de3f5bc-8761-491e-9a81-9a03e796ca1b-kube-api-access-djtmj\") pod \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.543813 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-inventory\") pod \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\" (UID: \"0de3f5bc-8761-491e-9a81-9a03e796ca1b\") " Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.551518 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0de3f5bc-8761-491e-9a81-9a03e796ca1b-kube-api-access-djtmj" (OuterVolumeSpecName: "kube-api-access-djtmj") pod "0de3f5bc-8761-491e-9a81-9a03e796ca1b" (UID: "0de3f5bc-8761-491e-9a81-9a03e796ca1b"). InnerVolumeSpecName "kube-api-access-djtmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.575467 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0de3f5bc-8761-491e-9a81-9a03e796ca1b" (UID: "0de3f5bc-8761-491e-9a81-9a03e796ca1b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.578519 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-inventory" (OuterVolumeSpecName: "inventory") pod "0de3f5bc-8761-491e-9a81-9a03e796ca1b" (UID: "0de3f5bc-8761-491e-9a81-9a03e796ca1b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.646599 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.646637 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djtmj\" (UniqueName: \"kubernetes.io/projected/0de3f5bc-8761-491e-9a81-9a03e796ca1b-kube-api-access-djtmj\") on node \"crc\" DevicePath \"\"" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.646650 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0de3f5bc-8761-491e-9a81-9a03e796ca1b-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.957921 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" event={"ID":"0de3f5bc-8761-491e-9a81-9a03e796ca1b","Type":"ContainerDied","Data":"f68016a53c1b011a78572fbd55948351591243e2cd97a0b5cebacfd2a4d955f0"} Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.957995 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f68016a53c1b011a78572fbd55948351591243e2cd97a0b5cebacfd2a4d955f0" Mar 13 08:08:26 crc kubenswrapper[4876]: I0313 08:08:26.958083 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fwnht" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.194747 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5"] Mar 13 08:08:27 crc kubenswrapper[4876]: E0313 08:08:27.195404 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0de3f5bc-8761-491e-9a81-9a03e796ca1b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.195439 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0de3f5bc-8761-491e-9a81-9a03e796ca1b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 08:08:27 crc kubenswrapper[4876]: E0313 08:08:27.195457 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5452b7b8-0dbd-46f7-91d4-2321bcb62f5f" containerName="oc" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.195467 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5452b7b8-0dbd-46f7-91d4-2321bcb62f5f" containerName="oc" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.195769 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0de3f5bc-8761-491e-9a81-9a03e796ca1b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.195797 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5452b7b8-0dbd-46f7-91d4-2321bcb62f5f" containerName="oc" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.197473 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.201198 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.201695 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.201912 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.212851 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5"] Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.216868 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.259250 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.259543 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.259805 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtl9f\" (UniqueName: \"kubernetes.io/projected/5bd6dd59-7427-4e76-8570-0a334476406c-kube-api-access-jtl9f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.362179 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.362321 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtl9f\" (UniqueName: \"kubernetes.io/projected/5bd6dd59-7427-4e76-8570-0a334476406c-kube-api-access-jtl9f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.362403 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.367898 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.368453 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.389763 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtl9f\" (UniqueName: \"kubernetes.io/projected/5bd6dd59-7427-4e76-8570-0a334476406c-kube-api-access-jtl9f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:27 crc kubenswrapper[4876]: I0313 08:08:27.523908 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:08:28 crc kubenswrapper[4876]: I0313 08:08:28.039428 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-67gcf"] Mar 13 08:08:28 crc kubenswrapper[4876]: I0313 08:08:28.049455 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-67gcf"] Mar 13 08:08:28 crc kubenswrapper[4876]: I0313 08:08:28.086950 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:08:28 crc kubenswrapper[4876]: I0313 08:08:28.087517 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5"] Mar 13 08:08:28 crc kubenswrapper[4876]: I0313 08:08:28.980393 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" event={"ID":"5bd6dd59-7427-4e76-8570-0a334476406c","Type":"ContainerStarted","Data":"d719448b54a121af38bfb8f7e6e99fbbc5923c5b2b0ed229685bc51d872d53e0"} Mar 13 08:08:28 crc kubenswrapper[4876]: I0313 08:08:28.980758 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" event={"ID":"5bd6dd59-7427-4e76-8570-0a334476406c","Type":"ContainerStarted","Data":"5a07e8da291ebe1a1be06a5d499fbcb0f104883b80ef48ed1eac8d69cececa0f"} Mar 13 08:08:29 crc kubenswrapper[4876]: I0313 08:08:29.000182 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" podStartSLOduration=1.503381782 podStartE2EDuration="2.00016125s" podCreationTimestamp="2026-03-13 08:08:27 +0000 UTC" firstStartedPulling="2026-03-13 08:08:28.08667205 +0000 UTC m=+1767.757451032" lastFinishedPulling="2026-03-13 08:08:28.583451518 +0000 UTC m=+1768.254230500" observedRunningTime="2026-03-13 08:08:28.998493575 +0000 UTC m=+1768.669272567" watchObservedRunningTime="2026-03-13 08:08:29.00016125 +0000 UTC m=+1768.670940222" Mar 13 08:08:29 crc kubenswrapper[4876]: I0313 08:08:29.052597 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b06a84e-50b9-4e3f-b57e-4db1b023699d" path="/var/lib/kubelet/pods/0b06a84e-50b9-4e3f-b57e-4db1b023699d/volumes" Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.057919 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c123-account-create-update-r29rm"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.060445 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5253-account-create-update-zqf6n"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.070472 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a6be-account-create-update-z9kh5"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.084472 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c123-account-create-update-r29rm"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.095433 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5253-account-create-update-zqf6n"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.107370 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a6be-account-create-update-z9kh5"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.115491 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-jm9nf"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.124031 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-xwt9g"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.133499 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-jm9nf"] Mar 13 08:08:33 crc kubenswrapper[4876]: I0313 08:08:33.141205 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-xwt9g"] Mar 13 08:08:35 crc kubenswrapper[4876]: I0313 08:08:35.047465 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27f72293-3842-46ab-87ad-fe2345ea4b3a" path="/var/lib/kubelet/pods/27f72293-3842-46ab-87ad-fe2345ea4b3a/volumes" Mar 13 08:08:35 crc kubenswrapper[4876]: I0313 08:08:35.048449 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44f174ba-30f6-4440-9d1c-037d413b5f90" path="/var/lib/kubelet/pods/44f174ba-30f6-4440-9d1c-037d413b5f90/volumes" Mar 13 08:08:35 crc kubenswrapper[4876]: I0313 08:08:35.048984 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ccc93a4-d813-4a7e-965d-2fdce73b92b5" path="/var/lib/kubelet/pods/4ccc93a4-d813-4a7e-965d-2fdce73b92b5/volumes" Mar 13 08:08:35 crc kubenswrapper[4876]: I0313 08:08:35.049592 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52e28a17-18b9-44a3-beec-cf688f768519" path="/var/lib/kubelet/pods/52e28a17-18b9-44a3-beec-cf688f768519/volumes" Mar 13 08:08:35 crc kubenswrapper[4876]: I0313 08:08:35.050694 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e16d2f43-d740-4edb-b41d-806e8bb28ff0" path="/var/lib/kubelet/pods/e16d2f43-d740-4edb-b41d-806e8bb28ff0/volumes" Mar 13 08:08:36 crc kubenswrapper[4876]: I0313 08:08:36.036887 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:08:36 crc kubenswrapper[4876]: E0313 08:08:36.037483 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:08:38 crc kubenswrapper[4876]: I0313 08:08:38.045372 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-45rb2"] Mar 13 08:08:38 crc kubenswrapper[4876]: I0313 08:08:38.059147 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-45rb2"] Mar 13 08:08:39 crc kubenswrapper[4876]: I0313 08:08:39.049420 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8523dd40-992c-4540-8bb0-2464da42a16c" path="/var/lib/kubelet/pods/8523dd40-992c-4540-8bb0-2464da42a16c/volumes" Mar 13 08:08:51 crc kubenswrapper[4876]: I0313 08:08:51.041585 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:08:51 crc kubenswrapper[4876]: E0313 08:08:51.042566 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:09:04 crc kubenswrapper[4876]: I0313 08:09:04.036151 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:09:04 crc kubenswrapper[4876]: E0313 08:09:04.037068 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:09:09 crc kubenswrapper[4876]: I0313 08:09:09.051323 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-hsm7j"] Mar 13 08:09:09 crc kubenswrapper[4876]: I0313 08:09:09.052517 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-hsm7j"] Mar 13 08:09:11 crc kubenswrapper[4876]: I0313 08:09:11.065035 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0b8464c-522b-44cc-8943-0f0770c2310b" path="/var/lib/kubelet/pods/e0b8464c-522b-44cc-8943-0f0770c2310b/volumes" Mar 13 08:09:15 crc kubenswrapper[4876]: I0313 08:09:15.036315 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:09:15 crc kubenswrapper[4876]: E0313 08:09:15.037103 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:09:20 crc kubenswrapper[4876]: I0313 08:09:20.723161 4876 scope.go:117] "RemoveContainer" containerID="4cec8ecf0ea3abac126fdd5020118cfd79550c900e56214461c4e3705c9b58ce" Mar 13 08:09:20 crc kubenswrapper[4876]: I0313 08:09:20.770031 4876 scope.go:117] "RemoveContainer" containerID="21004013435ebbea8d503cc46964a14804d066956a4b11aab90d7cba67ca1b1c" Mar 13 08:09:20 crc kubenswrapper[4876]: I0313 08:09:20.810716 4876 scope.go:117] "RemoveContainer" containerID="9b8cd85de9248213a81da8270122abfd4dca68b5f7cb4c51879f0ae973ef3117" Mar 13 08:09:20 crc kubenswrapper[4876]: I0313 08:09:20.850367 4876 scope.go:117] "RemoveContainer" containerID="b85d91a2a04c2f2f36ad7d8b2e07b84b66502458e73bb82449707d5930ae404f" Mar 13 08:09:20 crc kubenswrapper[4876]: I0313 08:09:20.897296 4876 scope.go:117] "RemoveContainer" containerID="79e044258014472c82a00a13c246a0d4373b7396c93188f36a1c7efe5d2fb7bd" Mar 13 08:09:20 crc kubenswrapper[4876]: I0313 08:09:20.941405 4876 scope.go:117] "RemoveContainer" containerID="50450a1e0da546a7f2ec5efe72a2a4ce3d0d9154e8b9b05e588df635f5db89ac" Mar 13 08:09:20 crc kubenswrapper[4876]: I0313 08:09:20.999734 4876 scope.go:117] "RemoveContainer" containerID="bbe1a539a48c93a928b82c30e621bf5e872a6f21d864308820fdfefc31cecbc0" Mar 13 08:09:21 crc kubenswrapper[4876]: I0313 08:09:21.033028 4876 scope.go:117] "RemoveContainer" containerID="6b137d93b7f9764555587381414a2363185119a97a017b13e0c0a831c8d99cd8" Mar 13 08:09:21 crc kubenswrapper[4876]: I0313 08:09:21.063015 4876 scope.go:117] "RemoveContainer" containerID="6c9ccd7f9ccc1773a10fad2983c6fae777e122684bccaaa3d24a91f678f5ce2a" Mar 13 08:09:21 crc kubenswrapper[4876]: I0313 08:09:21.109294 4876 scope.go:117] "RemoveContainer" containerID="943035bd6fadc99ace78e87aa1087aa1f4c16b922ebe27ec8f1ff69c3e4d5a5e" Mar 13 08:09:21 crc kubenswrapper[4876]: I0313 08:09:21.131981 4876 scope.go:117] "RemoveContainer" containerID="5d3759610c114fd90a7e5012fd273548fce3da5c843f4daf17c879cb01455d7d" Mar 13 08:09:23 crc kubenswrapper[4876]: I0313 08:09:23.070332 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dtrx5"] Mar 13 08:09:23 crc kubenswrapper[4876]: I0313 08:09:23.081636 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dtrx5"] Mar 13 08:09:25 crc kubenswrapper[4876]: I0313 08:09:25.047756 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88446f52-6434-4754-b225-3e79cbccac1b" path="/var/lib/kubelet/pods/88446f52-6434-4754-b225-3e79cbccac1b/volumes" Mar 13 08:09:26 crc kubenswrapper[4876]: I0313 08:09:26.035866 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:09:26 crc kubenswrapper[4876]: E0313 08:09:26.036498 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:09:29 crc kubenswrapper[4876]: I0313 08:09:29.047198 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-l75wd"] Mar 13 08:09:29 crc kubenswrapper[4876]: I0313 08:09:29.047618 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-l75wd"] Mar 13 08:09:31 crc kubenswrapper[4876]: I0313 08:09:31.054925 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="669de0e1-d1a3-4298-bee2-e3713fd7f652" path="/var/lib/kubelet/pods/669de0e1-d1a3-4298-bee2-e3713fd7f652/volumes" Mar 13 08:09:34 crc kubenswrapper[4876]: I0313 08:09:34.733372 4876 generic.go:334] "Generic (PLEG): container finished" podID="5bd6dd59-7427-4e76-8570-0a334476406c" containerID="d719448b54a121af38bfb8f7e6e99fbbc5923c5b2b0ed229685bc51d872d53e0" exitCode=0 Mar 13 08:09:34 crc kubenswrapper[4876]: I0313 08:09:34.733469 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" event={"ID":"5bd6dd59-7427-4e76-8570-0a334476406c","Type":"ContainerDied","Data":"d719448b54a121af38bfb8f7e6e99fbbc5923c5b2b0ed229685bc51d872d53e0"} Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.220584 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.295869 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-ssh-key-openstack-edpm-ipam\") pod \"5bd6dd59-7427-4e76-8570-0a334476406c\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.296108 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtl9f\" (UniqueName: \"kubernetes.io/projected/5bd6dd59-7427-4e76-8570-0a334476406c-kube-api-access-jtl9f\") pod \"5bd6dd59-7427-4e76-8570-0a334476406c\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.296163 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-inventory\") pod \"5bd6dd59-7427-4e76-8570-0a334476406c\" (UID: \"5bd6dd59-7427-4e76-8570-0a334476406c\") " Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.311310 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bd6dd59-7427-4e76-8570-0a334476406c-kube-api-access-jtl9f" (OuterVolumeSpecName: "kube-api-access-jtl9f") pod "5bd6dd59-7427-4e76-8570-0a334476406c" (UID: "5bd6dd59-7427-4e76-8570-0a334476406c"). InnerVolumeSpecName "kube-api-access-jtl9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.328822 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-inventory" (OuterVolumeSpecName: "inventory") pod "5bd6dd59-7427-4e76-8570-0a334476406c" (UID: "5bd6dd59-7427-4e76-8570-0a334476406c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.329354 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5bd6dd59-7427-4e76-8570-0a334476406c" (UID: "5bd6dd59-7427-4e76-8570-0a334476406c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.398946 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.398992 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtl9f\" (UniqueName: \"kubernetes.io/projected/5bd6dd59-7427-4e76-8570-0a334476406c-kube-api-access-jtl9f\") on node \"crc\" DevicePath \"\"" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.399007 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bd6dd59-7427-4e76-8570-0a334476406c-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.753841 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" event={"ID":"5bd6dd59-7427-4e76-8570-0a334476406c","Type":"ContainerDied","Data":"5a07e8da291ebe1a1be06a5d499fbcb0f104883b80ef48ed1eac8d69cececa0f"} Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.753911 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a07e8da291ebe1a1be06a5d499fbcb0f104883b80ef48ed1eac8d69cececa0f" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.753921 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.842779 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp"] Mar 13 08:09:36 crc kubenswrapper[4876]: E0313 08:09:36.843402 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd6dd59-7427-4e76-8570-0a334476406c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.843428 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd6dd59-7427-4e76-8570-0a334476406c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.844912 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd6dd59-7427-4e76-8570-0a334476406c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.846027 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.848426 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.848769 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.848942 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.849686 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.855366 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp"] Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.909722 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.909798 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdnlz\" (UniqueName: \"kubernetes.io/projected/5b3fa1e6-19bc-4852-bf70-69051aa084d6-kube-api-access-vdnlz\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:36 crc kubenswrapper[4876]: I0313 08:09:36.909899 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.011423 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.011582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.011624 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdnlz\" (UniqueName: \"kubernetes.io/projected/5b3fa1e6-19bc-4852-bf70-69051aa084d6-kube-api-access-vdnlz\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.021423 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.023404 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.034164 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdnlz\" (UniqueName: \"kubernetes.io/projected/5b3fa1e6-19bc-4852-bf70-69051aa084d6-kube-api-access-vdnlz\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qljdp\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.169409 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.726602 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp"] Mar 13 08:09:37 crc kubenswrapper[4876]: I0313 08:09:37.764758 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" event={"ID":"5b3fa1e6-19bc-4852-bf70-69051aa084d6","Type":"ContainerStarted","Data":"f99973705d90e811face6e6dcd2c7448c145922e79451796e54d29063081fbe7"} Mar 13 08:09:38 crc kubenswrapper[4876]: I0313 08:09:38.774763 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" event={"ID":"5b3fa1e6-19bc-4852-bf70-69051aa084d6","Type":"ContainerStarted","Data":"35b129598d240404618dd24bf5c0582f5ad30541df7ff4c2d68aa2905697f712"} Mar 13 08:09:38 crc kubenswrapper[4876]: I0313 08:09:38.801460 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" podStartSLOduration=2.337453245 podStartE2EDuration="2.801430224s" podCreationTimestamp="2026-03-13 08:09:36 +0000 UTC" firstStartedPulling="2026-03-13 08:09:37.732116012 +0000 UTC m=+1837.402894994" lastFinishedPulling="2026-03-13 08:09:38.196092991 +0000 UTC m=+1837.866871973" observedRunningTime="2026-03-13 08:09:38.79287276 +0000 UTC m=+1838.463651742" watchObservedRunningTime="2026-03-13 08:09:38.801430224 +0000 UTC m=+1838.472209206" Mar 13 08:09:39 crc kubenswrapper[4876]: I0313 08:09:39.034293 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-rzbxc"] Mar 13 08:09:39 crc kubenswrapper[4876]: I0313 08:09:39.048137 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-ltwdh"] Mar 13 08:09:39 crc kubenswrapper[4876]: I0313 08:09:39.049553 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:09:39 crc kubenswrapper[4876]: E0313 08:09:39.049897 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:09:39 crc kubenswrapper[4876]: I0313 08:09:39.054605 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-rzbxc"] Mar 13 08:09:39 crc kubenswrapper[4876]: I0313 08:09:39.065522 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-ltwdh"] Mar 13 08:09:41 crc kubenswrapper[4876]: I0313 08:09:41.050794 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d42486d-85b1-4384-91fc-10552a2c404b" path="/var/lib/kubelet/pods/5d42486d-85b1-4384-91fc-10552a2c404b/volumes" Mar 13 08:09:41 crc kubenswrapper[4876]: I0313 08:09:41.052198 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86d9e4f3-c1e8-40fe-a6b8-79507f63ca32" path="/var/lib/kubelet/pods/86d9e4f3-c1e8-40fe-a6b8-79507f63ca32/volumes" Mar 13 08:09:43 crc kubenswrapper[4876]: I0313 08:09:43.824370 4876 generic.go:334] "Generic (PLEG): container finished" podID="5b3fa1e6-19bc-4852-bf70-69051aa084d6" containerID="35b129598d240404618dd24bf5c0582f5ad30541df7ff4c2d68aa2905697f712" exitCode=0 Mar 13 08:09:43 crc kubenswrapper[4876]: I0313 08:09:43.824409 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" event={"ID":"5b3fa1e6-19bc-4852-bf70-69051aa084d6","Type":"ContainerDied","Data":"35b129598d240404618dd24bf5c0582f5ad30541df7ff4c2d68aa2905697f712"} Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.273075 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.418412 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdnlz\" (UniqueName: \"kubernetes.io/projected/5b3fa1e6-19bc-4852-bf70-69051aa084d6-kube-api-access-vdnlz\") pod \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.418626 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-ssh-key-openstack-edpm-ipam\") pod \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.418692 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-inventory\") pod \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\" (UID: \"5b3fa1e6-19bc-4852-bf70-69051aa084d6\") " Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.425620 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b3fa1e6-19bc-4852-bf70-69051aa084d6-kube-api-access-vdnlz" (OuterVolumeSpecName: "kube-api-access-vdnlz") pod "5b3fa1e6-19bc-4852-bf70-69051aa084d6" (UID: "5b3fa1e6-19bc-4852-bf70-69051aa084d6"). InnerVolumeSpecName "kube-api-access-vdnlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.454585 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-inventory" (OuterVolumeSpecName: "inventory") pod "5b3fa1e6-19bc-4852-bf70-69051aa084d6" (UID: "5b3fa1e6-19bc-4852-bf70-69051aa084d6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.455515 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5b3fa1e6-19bc-4852-bf70-69051aa084d6" (UID: "5b3fa1e6-19bc-4852-bf70-69051aa084d6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.521198 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdnlz\" (UniqueName: \"kubernetes.io/projected/5b3fa1e6-19bc-4852-bf70-69051aa084d6-kube-api-access-vdnlz\") on node \"crc\" DevicePath \"\"" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.521255 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.521272 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b3fa1e6-19bc-4852-bf70-69051aa084d6-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.848259 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" event={"ID":"5b3fa1e6-19bc-4852-bf70-69051aa084d6","Type":"ContainerDied","Data":"f99973705d90e811face6e6dcd2c7448c145922e79451796e54d29063081fbe7"} Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.848324 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f99973705d90e811face6e6dcd2c7448c145922e79451796e54d29063081fbe7" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.848750 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qljdp" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.923958 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw"] Mar 13 08:09:45 crc kubenswrapper[4876]: E0313 08:09:45.924410 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3fa1e6-19bc-4852-bf70-69051aa084d6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.924430 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3fa1e6-19bc-4852-bf70-69051aa084d6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.924667 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3fa1e6-19bc-4852-bf70-69051aa084d6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.925438 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.928347 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.928640 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.928940 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.929717 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:09:45 crc kubenswrapper[4876]: I0313 08:09:45.936334 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw"] Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.030762 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.030838 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.030930 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc8pt\" (UniqueName: \"kubernetes.io/projected/fb7b397a-4ce3-4a83-b236-5baac12e2a15-kube-api-access-jc8pt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.132558 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc8pt\" (UniqueName: \"kubernetes.io/projected/fb7b397a-4ce3-4a83-b236-5baac12e2a15-kube-api-access-jc8pt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.132731 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.132790 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.137734 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.138255 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.157983 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc8pt\" (UniqueName: \"kubernetes.io/projected/fb7b397a-4ce3-4a83-b236-5baac12e2a15-kube-api-access-jc8pt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4m5cw\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.254566 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.810781 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw"] Mar 13 08:09:46 crc kubenswrapper[4876]: I0313 08:09:46.860453 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" event={"ID":"fb7b397a-4ce3-4a83-b236-5baac12e2a15","Type":"ContainerStarted","Data":"9fbb1a9f4180c96afe9be4086299046d1c081ff1f72e6bd4c0e39f9b6b195bbe"} Mar 13 08:09:47 crc kubenswrapper[4876]: I0313 08:09:47.873800 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" event={"ID":"fb7b397a-4ce3-4a83-b236-5baac12e2a15","Type":"ContainerStarted","Data":"c40d47c0b6898e2983574deee5ff6f027db5d83bac74104ba722c782a3f36f18"} Mar 13 08:09:47 crc kubenswrapper[4876]: I0313 08:09:47.927662 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" podStartSLOduration=2.496633365 podStartE2EDuration="2.927634479s" podCreationTimestamp="2026-03-13 08:09:45 +0000 UTC" firstStartedPulling="2026-03-13 08:09:46.819338649 +0000 UTC m=+1846.490117631" lastFinishedPulling="2026-03-13 08:09:47.250339763 +0000 UTC m=+1846.921118745" observedRunningTime="2026-03-13 08:09:47.896130276 +0000 UTC m=+1847.566909268" watchObservedRunningTime="2026-03-13 08:09:47.927634479 +0000 UTC m=+1847.598413461" Mar 13 08:09:51 crc kubenswrapper[4876]: I0313 08:09:51.045156 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:09:51 crc kubenswrapper[4876]: E0313 08:09:51.045806 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.160926 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556490-ghmfx"] Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.163189 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556490-ghmfx" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.166641 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.166771 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.167050 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.175720 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556490-ghmfx"] Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.225124 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7b4d\" (UniqueName: \"kubernetes.io/projected/afc332f0-d43a-4853-8017-47f67a849f2f-kube-api-access-x7b4d\") pod \"auto-csr-approver-29556490-ghmfx\" (UID: \"afc332f0-d43a-4853-8017-47f67a849f2f\") " pod="openshift-infra/auto-csr-approver-29556490-ghmfx" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.327885 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7b4d\" (UniqueName: \"kubernetes.io/projected/afc332f0-d43a-4853-8017-47f67a849f2f-kube-api-access-x7b4d\") pod \"auto-csr-approver-29556490-ghmfx\" (UID: \"afc332f0-d43a-4853-8017-47f67a849f2f\") " pod="openshift-infra/auto-csr-approver-29556490-ghmfx" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.347637 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7b4d\" (UniqueName: \"kubernetes.io/projected/afc332f0-d43a-4853-8017-47f67a849f2f-kube-api-access-x7b4d\") pod \"auto-csr-approver-29556490-ghmfx\" (UID: \"afc332f0-d43a-4853-8017-47f67a849f2f\") " pod="openshift-infra/auto-csr-approver-29556490-ghmfx" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.493693 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556490-ghmfx" Mar 13 08:10:00 crc kubenswrapper[4876]: I0313 08:10:00.972327 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556490-ghmfx"] Mar 13 08:10:00 crc kubenswrapper[4876]: W0313 08:10:00.992337 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafc332f0_d43a_4853_8017_47f67a849f2f.slice/crio-42c8ecbbe371921d617c656f7e15835c9d0dc61d50faac480561925af42de207 WatchSource:0}: Error finding container 42c8ecbbe371921d617c656f7e15835c9d0dc61d50faac480561925af42de207: Status 404 returned error can't find the container with id 42c8ecbbe371921d617c656f7e15835c9d0dc61d50faac480561925af42de207 Mar 13 08:10:02 crc kubenswrapper[4876]: I0313 08:10:02.012312 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556490-ghmfx" event={"ID":"afc332f0-d43a-4853-8017-47f67a849f2f","Type":"ContainerStarted","Data":"42c8ecbbe371921d617c656f7e15835c9d0dc61d50faac480561925af42de207"} Mar 13 08:10:03 crc kubenswrapper[4876]: I0313 08:10:03.029663 4876 generic.go:334] "Generic (PLEG): container finished" podID="afc332f0-d43a-4853-8017-47f67a849f2f" containerID="24cf542e7878ea6641223e58c113d0325a2281dfc91432f3ea4f68153daf7050" exitCode=0 Mar 13 08:10:03 crc kubenswrapper[4876]: I0313 08:10:03.029779 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556490-ghmfx" event={"ID":"afc332f0-d43a-4853-8017-47f67a849f2f","Type":"ContainerDied","Data":"24cf542e7878ea6641223e58c113d0325a2281dfc91432f3ea4f68153daf7050"} Mar 13 08:10:04 crc kubenswrapper[4876]: I0313 08:10:04.446520 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556490-ghmfx" Mar 13 08:10:04 crc kubenswrapper[4876]: I0313 08:10:04.520372 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7b4d\" (UniqueName: \"kubernetes.io/projected/afc332f0-d43a-4853-8017-47f67a849f2f-kube-api-access-x7b4d\") pod \"afc332f0-d43a-4853-8017-47f67a849f2f\" (UID: \"afc332f0-d43a-4853-8017-47f67a849f2f\") " Mar 13 08:10:04 crc kubenswrapper[4876]: I0313 08:10:04.529759 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc332f0-d43a-4853-8017-47f67a849f2f-kube-api-access-x7b4d" (OuterVolumeSpecName: "kube-api-access-x7b4d") pod "afc332f0-d43a-4853-8017-47f67a849f2f" (UID: "afc332f0-d43a-4853-8017-47f67a849f2f"). InnerVolumeSpecName "kube-api-access-x7b4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:10:04 crc kubenswrapper[4876]: I0313 08:10:04.623406 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7b4d\" (UniqueName: \"kubernetes.io/projected/afc332f0-d43a-4853-8017-47f67a849f2f-kube-api-access-x7b4d\") on node \"crc\" DevicePath \"\"" Mar 13 08:10:05 crc kubenswrapper[4876]: I0313 08:10:05.051874 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556490-ghmfx" event={"ID":"afc332f0-d43a-4853-8017-47f67a849f2f","Type":"ContainerDied","Data":"42c8ecbbe371921d617c656f7e15835c9d0dc61d50faac480561925af42de207"} Mar 13 08:10:05 crc kubenswrapper[4876]: I0313 08:10:05.051929 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42c8ecbbe371921d617c656f7e15835c9d0dc61d50faac480561925af42de207" Mar 13 08:10:05 crc kubenswrapper[4876]: I0313 08:10:05.051961 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556490-ghmfx" Mar 13 08:10:05 crc kubenswrapper[4876]: I0313 08:10:05.550933 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556484-mlmsn"] Mar 13 08:10:05 crc kubenswrapper[4876]: I0313 08:10:05.561370 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556484-mlmsn"] Mar 13 08:10:06 crc kubenswrapper[4876]: I0313 08:10:06.035574 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:10:06 crc kubenswrapper[4876]: E0313 08:10:06.036265 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:10:07 crc kubenswrapper[4876]: I0313 08:10:07.050651 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe1c50d-4a24-479d-9d08-780fc1c0e0a7" path="/var/lib/kubelet/pods/7fe1c50d-4a24-479d-9d08-780fc1c0e0a7/volumes" Mar 13 08:10:20 crc kubenswrapper[4876]: I0313 08:10:20.035374 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:10:20 crc kubenswrapper[4876]: E0313 08:10:20.036303 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:10:21 crc kubenswrapper[4876]: I0313 08:10:21.376530 4876 scope.go:117] "RemoveContainer" containerID="fe3da91ed9a3a3ca316b34ea0f7c95a6a1766d17e9327d1fe2218db9cab2dd6c" Mar 13 08:10:21 crc kubenswrapper[4876]: I0313 08:10:21.433968 4876 scope.go:117] "RemoveContainer" containerID="396ac8d43845968d60c9c9a325f780eaa34e76d7d1cb79e591f45a4d0cb927d7" Mar 13 08:10:21 crc kubenswrapper[4876]: I0313 08:10:21.481583 4876 scope.go:117] "RemoveContainer" containerID="33ad2db5455f1f79c44d2ce36ec5d0ed2b28712faf211c4126b8ffdd05dca8d7" Mar 13 08:10:21 crc kubenswrapper[4876]: I0313 08:10:21.531129 4876 scope.go:117] "RemoveContainer" containerID="ef20163b689236111fb8e6fc380a33a061900a45cc3f2e63ffa286b5d0215ea7" Mar 13 08:10:21 crc kubenswrapper[4876]: I0313 08:10:21.602940 4876 scope.go:117] "RemoveContainer" containerID="849536096db85d5fccb4972e93f7f2407fb4aea0e532aaed84997018aa2c3d03" Mar 13 08:10:22 crc kubenswrapper[4876]: I0313 08:10:22.049056 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-x9tx5"] Mar 13 08:10:22 crc kubenswrapper[4876]: I0313 08:10:22.062100 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-3c87-account-create-update-mgw9g"] Mar 13 08:10:22 crc kubenswrapper[4876]: I0313 08:10:22.074041 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-x9tx5"] Mar 13 08:10:22 crc kubenswrapper[4876]: I0313 08:10:22.085729 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-3c87-account-create-update-mgw9g"] Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.056426 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b8e947-a71d-4365-a028-7f8d6d184577" path="/var/lib/kubelet/pods/51b8e947-a71d-4365-a028-7f8d6d184577/volumes" Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.057531 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8229f43e-63f9-41a0-8197-4af96c27ee48" path="/var/lib/kubelet/pods/8229f43e-63f9-41a0-8197-4af96c27ee48/volumes" Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.058134 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-kl96z"] Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.059052 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-wxvxg"] Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.072306 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-wxvxg"] Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.130945 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-75fc-account-create-update-zjt7n"] Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.165582 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-kl96z"] Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.178733 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-102d-account-create-update-kvlpr"] Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.188264 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-75fc-account-create-update-zjt7n"] Mar 13 08:10:23 crc kubenswrapper[4876]: I0313 08:10:23.195485 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-102d-account-create-update-kvlpr"] Mar 13 08:10:24 crc kubenswrapper[4876]: I0313 08:10:24.269015 4876 generic.go:334] "Generic (PLEG): container finished" podID="fb7b397a-4ce3-4a83-b236-5baac12e2a15" containerID="c40d47c0b6898e2983574deee5ff6f027db5d83bac74104ba722c782a3f36f18" exitCode=0 Mar 13 08:10:24 crc kubenswrapper[4876]: I0313 08:10:24.269098 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" event={"ID":"fb7b397a-4ce3-4a83-b236-5baac12e2a15","Type":"ContainerDied","Data":"c40d47c0b6898e2983574deee5ff6f027db5d83bac74104ba722c782a3f36f18"} Mar 13 08:10:25 crc kubenswrapper[4876]: I0313 08:10:25.046848 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fec4b3c-6d95-4ed5-ac57-40da620b201c" path="/var/lib/kubelet/pods/3fec4b3c-6d95-4ed5-ac57-40da620b201c/volumes" Mar 13 08:10:25 crc kubenswrapper[4876]: I0313 08:10:25.047451 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c8d7466-ceca-4879-9263-9c9eeb0d3a7a" path="/var/lib/kubelet/pods/5c8d7466-ceca-4879-9263-9c9eeb0d3a7a/volumes" Mar 13 08:10:25 crc kubenswrapper[4876]: I0313 08:10:25.047999 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0939f9-b421-4707-861c-b43fae99f079" path="/var/lib/kubelet/pods/6e0939f9-b421-4707-861c-b43fae99f079/volumes" Mar 13 08:10:25 crc kubenswrapper[4876]: I0313 08:10:25.048584 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ecb0fb1-66c7-400e-90cb-7f5bbe847f12" path="/var/lib/kubelet/pods/7ecb0fb1-66c7-400e-90cb-7f5bbe847f12/volumes" Mar 13 08:10:25 crc kubenswrapper[4876]: I0313 08:10:25.924724 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.045151 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc8pt\" (UniqueName: \"kubernetes.io/projected/fb7b397a-4ce3-4a83-b236-5baac12e2a15-kube-api-access-jc8pt\") pod \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.045312 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-ssh-key-openstack-edpm-ipam\") pod \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.045420 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-inventory\") pod \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\" (UID: \"fb7b397a-4ce3-4a83-b236-5baac12e2a15\") " Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.057171 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb7b397a-4ce3-4a83-b236-5baac12e2a15-kube-api-access-jc8pt" (OuterVolumeSpecName: "kube-api-access-jc8pt") pod "fb7b397a-4ce3-4a83-b236-5baac12e2a15" (UID: "fb7b397a-4ce3-4a83-b236-5baac12e2a15"). InnerVolumeSpecName "kube-api-access-jc8pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.073959 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-inventory" (OuterVolumeSpecName: "inventory") pod "fb7b397a-4ce3-4a83-b236-5baac12e2a15" (UID: "fb7b397a-4ce3-4a83-b236-5baac12e2a15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.075996 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fb7b397a-4ce3-4a83-b236-5baac12e2a15" (UID: "fb7b397a-4ce3-4a83-b236-5baac12e2a15"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.148035 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc8pt\" (UniqueName: \"kubernetes.io/projected/fb7b397a-4ce3-4a83-b236-5baac12e2a15-kube-api-access-jc8pt\") on node \"crc\" DevicePath \"\"" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.148075 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.148089 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb7b397a-4ce3-4a83-b236-5baac12e2a15-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.291737 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" event={"ID":"fb7b397a-4ce3-4a83-b236-5baac12e2a15","Type":"ContainerDied","Data":"9fbb1a9f4180c96afe9be4086299046d1c081ff1f72e6bd4c0e39f9b6b195bbe"} Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.291797 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fbb1a9f4180c96afe9be4086299046d1c081ff1f72e6bd4c0e39f9b6b195bbe" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.291864 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4m5cw" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.376305 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4"] Mar 13 08:10:26 crc kubenswrapper[4876]: E0313 08:10:26.376778 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc332f0-d43a-4853-8017-47f67a849f2f" containerName="oc" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.376807 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc332f0-d43a-4853-8017-47f67a849f2f" containerName="oc" Mar 13 08:10:26 crc kubenswrapper[4876]: E0313 08:10:26.376845 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7b397a-4ce3-4a83-b236-5baac12e2a15" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.376854 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7b397a-4ce3-4a83-b236-5baac12e2a15" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.377098 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7b397a-4ce3-4a83-b236-5baac12e2a15" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.377123 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc332f0-d43a-4853-8017-47f67a849f2f" containerName="oc" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.377816 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.380752 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.381267 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.381555 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.384603 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.386898 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4"] Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.453641 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69mj4\" (UniqueName: \"kubernetes.io/projected/5268b132-b205-4741-8680-781f73f3209a-kube-api-access-69mj4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.453716 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.453808 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.555627 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69mj4\" (UniqueName: \"kubernetes.io/projected/5268b132-b205-4741-8680-781f73f3209a-kube-api-access-69mj4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.556631 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.557583 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.560726 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.562819 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.575978 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69mj4\" (UniqueName: \"kubernetes.io/projected/5268b132-b205-4741-8680-781f73f3209a-kube-api-access-69mj4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:26 crc kubenswrapper[4876]: I0313 08:10:26.702782 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:10:27 crc kubenswrapper[4876]: I0313 08:10:27.330224 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4"] Mar 13 08:10:28 crc kubenswrapper[4876]: I0313 08:10:28.319941 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" event={"ID":"5268b132-b205-4741-8680-781f73f3209a","Type":"ContainerStarted","Data":"f9eadefe9d3ddfcc1c1cc4940ad68fac084419a8924d11b8ecd4cb09f68bc47a"} Mar 13 08:10:28 crc kubenswrapper[4876]: I0313 08:10:28.320355 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" event={"ID":"5268b132-b205-4741-8680-781f73f3209a","Type":"ContainerStarted","Data":"6eed2755384e79cf9ab6e15d9dccd84560d156ce543e31051710a815a91edcf3"} Mar 13 08:10:28 crc kubenswrapper[4876]: I0313 08:10:28.348049 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" podStartSLOduration=1.894491567 podStartE2EDuration="2.348018836s" podCreationTimestamp="2026-03-13 08:10:26 +0000 UTC" firstStartedPulling="2026-03-13 08:10:27.337430375 +0000 UTC m=+1887.008209347" lastFinishedPulling="2026-03-13 08:10:27.790957634 +0000 UTC m=+1887.461736616" observedRunningTime="2026-03-13 08:10:28.338696288 +0000 UTC m=+1888.009475270" watchObservedRunningTime="2026-03-13 08:10:28.348018836 +0000 UTC m=+1888.018797818" Mar 13 08:10:33 crc kubenswrapper[4876]: I0313 08:10:33.040690 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:10:33 crc kubenswrapper[4876]: E0313 08:10:33.041846 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:10:45 crc kubenswrapper[4876]: I0313 08:10:45.036028 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:10:45 crc kubenswrapper[4876]: E0313 08:10:45.036742 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:10:53 crc kubenswrapper[4876]: I0313 08:10:53.055047 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzp92"] Mar 13 08:10:53 crc kubenswrapper[4876]: I0313 08:10:53.068266 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jzp92"] Mar 13 08:10:55 crc kubenswrapper[4876]: I0313 08:10:55.047968 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce" path="/var/lib/kubelet/pods/d934fb45-0f9f-490d-9ba1-8a2bd5a8dfce/volumes" Mar 13 08:10:56 crc kubenswrapper[4876]: I0313 08:10:56.036280 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:10:56 crc kubenswrapper[4876]: I0313 08:10:56.611975 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"d0c73df6ed9f0626ecd967ee4e119fdb9bc95174ba5b98bf45405cb2d0c94943"} Mar 13 08:11:15 crc kubenswrapper[4876]: I0313 08:11:15.052341 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5st8q"] Mar 13 08:11:15 crc kubenswrapper[4876]: I0313 08:11:15.067885 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lrrb8"] Mar 13 08:11:15 crc kubenswrapper[4876]: I0313 08:11:15.078257 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lrrb8"] Mar 13 08:11:15 crc kubenswrapper[4876]: I0313 08:11:15.086073 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5st8q"] Mar 13 08:11:17 crc kubenswrapper[4876]: I0313 08:11:17.051658 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39f93996-aba4-43e8-aa87-2e0a2cd9c1d4" path="/var/lib/kubelet/pods/39f93996-aba4-43e8-aa87-2e0a2cd9c1d4/volumes" Mar 13 08:11:17 crc kubenswrapper[4876]: I0313 08:11:17.053758 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="812ab290-71fc-4db3-94af-6daffce9da81" path="/var/lib/kubelet/pods/812ab290-71fc-4db3-94af-6daffce9da81/volumes" Mar 13 08:11:20 crc kubenswrapper[4876]: I0313 08:11:20.894543 4876 generic.go:334] "Generic (PLEG): container finished" podID="5268b132-b205-4741-8680-781f73f3209a" containerID="f9eadefe9d3ddfcc1c1cc4940ad68fac084419a8924d11b8ecd4cb09f68bc47a" exitCode=0 Mar 13 08:11:20 crc kubenswrapper[4876]: I0313 08:11:20.894663 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" event={"ID":"5268b132-b205-4741-8680-781f73f3209a","Type":"ContainerDied","Data":"f9eadefe9d3ddfcc1c1cc4940ad68fac084419a8924d11b8ecd4cb09f68bc47a"} Mar 13 08:11:21 crc kubenswrapper[4876]: I0313 08:11:21.741743 4876 scope.go:117] "RemoveContainer" containerID="a0c2c9cde78e2ddbd12545a5e60d19d0702f4e03f8e339f2e316ec52a739d6c3" Mar 13 08:11:21 crc kubenswrapper[4876]: I0313 08:11:21.788726 4876 scope.go:117] "RemoveContainer" containerID="ea3460326a2d4e125ca36fabec1f6ab8865870c613830208ba3ed170b8678ced" Mar 13 08:11:21 crc kubenswrapper[4876]: I0313 08:11:21.846357 4876 scope.go:117] "RemoveContainer" containerID="d3e8eecbdc105c24b74b8b8b513b6288e98ca1cd156c41fa36e38a8cc9ef88d4" Mar 13 08:11:21 crc kubenswrapper[4876]: I0313 08:11:21.905126 4876 scope.go:117] "RemoveContainer" containerID="58ded0c1d939e2e1aaaa3bcfa7a860776ba705aba34f9551cc684c9b8cdaabd6" Mar 13 08:11:21 crc kubenswrapper[4876]: I0313 08:11:21.983465 4876 scope.go:117] "RemoveContainer" containerID="766fe4c0e9d9606bd4a06d79c54564e2fd59dcec56c37a379437bb0a4821939a" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.009292 4876 scope.go:117] "RemoveContainer" containerID="dd54696bdd240b84feca773d507f3bd64ce6cfddb93affa36fdfda799f0c1f0f" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.067497 4876 scope.go:117] "RemoveContainer" containerID="4401cf7ec14b2aba19ace026bf6e00c3b0a74f90c86153ad7eb9df1e8ffb6cb3" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.097797 4876 scope.go:117] "RemoveContainer" containerID="2a7835d82325345577bbddaa180c61e60521ec132ac9d86add5588dfc18edce0" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.167629 4876 scope.go:117] "RemoveContainer" containerID="b2c87db8f35c5801cef873bcdfa601e4673a38239f4e43e1199f1892dca44fe2" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.323509 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.429494 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-ssh-key-openstack-edpm-ipam\") pod \"5268b132-b205-4741-8680-781f73f3209a\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.429544 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-inventory\") pod \"5268b132-b205-4741-8680-781f73f3209a\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.429768 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69mj4\" (UniqueName: \"kubernetes.io/projected/5268b132-b205-4741-8680-781f73f3209a-kube-api-access-69mj4\") pod \"5268b132-b205-4741-8680-781f73f3209a\" (UID: \"5268b132-b205-4741-8680-781f73f3209a\") " Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.439204 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5268b132-b205-4741-8680-781f73f3209a-kube-api-access-69mj4" (OuterVolumeSpecName: "kube-api-access-69mj4") pod "5268b132-b205-4741-8680-781f73f3209a" (UID: "5268b132-b205-4741-8680-781f73f3209a"). InnerVolumeSpecName "kube-api-access-69mj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.464373 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5268b132-b205-4741-8680-781f73f3209a" (UID: "5268b132-b205-4741-8680-781f73f3209a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.468066 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-inventory" (OuterVolumeSpecName: "inventory") pod "5268b132-b205-4741-8680-781f73f3209a" (UID: "5268b132-b205-4741-8680-781f73f3209a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.533047 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.533469 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5268b132-b205-4741-8680-781f73f3209a-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.533565 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69mj4\" (UniqueName: \"kubernetes.io/projected/5268b132-b205-4741-8680-781f73f3209a-kube-api-access-69mj4\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.921162 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" event={"ID":"5268b132-b205-4741-8680-781f73f3209a","Type":"ContainerDied","Data":"6eed2755384e79cf9ab6e15d9dccd84560d156ce543e31051710a815a91edcf3"} Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.921212 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6eed2755384e79cf9ab6e15d9dccd84560d156ce543e31051710a815a91edcf3" Mar 13 08:11:22 crc kubenswrapper[4876]: I0313 08:11:22.921295 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.089444 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6q2l4"] Mar 13 08:11:23 crc kubenswrapper[4876]: E0313 08:11:23.090368 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5268b132-b205-4741-8680-781f73f3209a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.090395 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5268b132-b205-4741-8680-781f73f3209a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.090739 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5268b132-b205-4741-8680-781f73f3209a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.091783 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.093984 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.094069 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.094185 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.095316 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6q2l4"] Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.097553 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.284908 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.285370 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsmt9\" (UniqueName: \"kubernetes.io/projected/ca0807c6-78af-45cf-994d-85563d8089ac-kube-api-access-zsmt9\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.285779 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.388909 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsmt9\" (UniqueName: \"kubernetes.io/projected/ca0807c6-78af-45cf-994d-85563d8089ac-kube-api-access-zsmt9\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.389067 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.389131 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.397003 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.403137 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.411293 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsmt9\" (UniqueName: \"kubernetes.io/projected/ca0807c6-78af-45cf-994d-85563d8089ac-kube-api-access-zsmt9\") pod \"ssh-known-hosts-edpm-deployment-6q2l4\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:23 crc kubenswrapper[4876]: I0313 08:11:23.415467 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:24 crc kubenswrapper[4876]: I0313 08:11:24.026495 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-6q2l4"] Mar 13 08:11:24 crc kubenswrapper[4876]: I0313 08:11:24.940421 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" event={"ID":"ca0807c6-78af-45cf-994d-85563d8089ac","Type":"ContainerStarted","Data":"e2a0c8eaba0dd2cc6839de682ed417b6fa970d35aa17bcc30204bf04bd2ca2f2"} Mar 13 08:11:24 crc kubenswrapper[4876]: I0313 08:11:24.941477 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" event={"ID":"ca0807c6-78af-45cf-994d-85563d8089ac","Type":"ContainerStarted","Data":"fa58763b4935e815030c00c93463625656b01aa66ae45fa8f1117c1f500e6d5f"} Mar 13 08:11:32 crc kubenswrapper[4876]: I0313 08:11:32.038824 4876 generic.go:334] "Generic (PLEG): container finished" podID="ca0807c6-78af-45cf-994d-85563d8089ac" containerID="e2a0c8eaba0dd2cc6839de682ed417b6fa970d35aa17bcc30204bf04bd2ca2f2" exitCode=0 Mar 13 08:11:32 crc kubenswrapper[4876]: I0313 08:11:32.038897 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" event={"ID":"ca0807c6-78af-45cf-994d-85563d8089ac","Type":"ContainerDied","Data":"e2a0c8eaba0dd2cc6839de682ed417b6fa970d35aa17bcc30204bf04bd2ca2f2"} Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.524526 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.626147 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsmt9\" (UniqueName: \"kubernetes.io/projected/ca0807c6-78af-45cf-994d-85563d8089ac-kube-api-access-zsmt9\") pod \"ca0807c6-78af-45cf-994d-85563d8089ac\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.626615 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-ssh-key-openstack-edpm-ipam\") pod \"ca0807c6-78af-45cf-994d-85563d8089ac\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.626637 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-inventory-0\") pod \"ca0807c6-78af-45cf-994d-85563d8089ac\" (UID: \"ca0807c6-78af-45cf-994d-85563d8089ac\") " Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.634631 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca0807c6-78af-45cf-994d-85563d8089ac-kube-api-access-zsmt9" (OuterVolumeSpecName: "kube-api-access-zsmt9") pod "ca0807c6-78af-45cf-994d-85563d8089ac" (UID: "ca0807c6-78af-45cf-994d-85563d8089ac"). InnerVolumeSpecName "kube-api-access-zsmt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.659668 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "ca0807c6-78af-45cf-994d-85563d8089ac" (UID: "ca0807c6-78af-45cf-994d-85563d8089ac"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.659871 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ca0807c6-78af-45cf-994d-85563d8089ac" (UID: "ca0807c6-78af-45cf-994d-85563d8089ac"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.730608 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsmt9\" (UniqueName: \"kubernetes.io/projected/ca0807c6-78af-45cf-994d-85563d8089ac-kube-api-access-zsmt9\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.730679 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:33 crc kubenswrapper[4876]: I0313 08:11:33.730694 4876 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ca0807c6-78af-45cf-994d-85563d8089ac-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.116853 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" event={"ID":"ca0807c6-78af-45cf-994d-85563d8089ac","Type":"ContainerDied","Data":"fa58763b4935e815030c00c93463625656b01aa66ae45fa8f1117c1f500e6d5f"} Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.116916 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa58763b4935e815030c00c93463625656b01aa66ae45fa8f1117c1f500e6d5f" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.116985 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-6q2l4" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.164483 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl"] Mar 13 08:11:34 crc kubenswrapper[4876]: E0313 08:11:34.164904 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca0807c6-78af-45cf-994d-85563d8089ac" containerName="ssh-known-hosts-edpm-deployment" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.164923 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca0807c6-78af-45cf-994d-85563d8089ac" containerName="ssh-known-hosts-edpm-deployment" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.165126 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca0807c6-78af-45cf-994d-85563d8089ac" containerName="ssh-known-hosts-edpm-deployment" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.165914 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.173633 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.173735 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.173819 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.174029 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.180090 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl"] Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.351513 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz5k4\" (UniqueName: \"kubernetes.io/projected/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-kube-api-access-fz5k4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.351584 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.351759 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.454693 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.455464 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz5k4\" (UniqueName: \"kubernetes.io/projected/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-kube-api-access-fz5k4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.455702 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.459944 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.459992 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.477761 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz5k4\" (UniqueName: \"kubernetes.io/projected/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-kube-api-access-fz5k4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hj2tl\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.492631 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:34 crc kubenswrapper[4876]: I0313 08:11:34.860753 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl"] Mar 13 08:11:35 crc kubenswrapper[4876]: I0313 08:11:35.129736 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" event={"ID":"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1","Type":"ContainerStarted","Data":"a33688e70e457b7f4c2111925e84071709e17da0ef7744913008d7eca96c2032"} Mar 13 08:11:36 crc kubenswrapper[4876]: I0313 08:11:36.140189 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" event={"ID":"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1","Type":"ContainerStarted","Data":"f94a02a93179d22563fe0d2062b9bac6c6b6b8f51f84406751d56051543d1734"} Mar 13 08:11:36 crc kubenswrapper[4876]: I0313 08:11:36.164194 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" podStartSLOduration=1.5295750620000002 podStartE2EDuration="2.164160617s" podCreationTimestamp="2026-03-13 08:11:34 +0000 UTC" firstStartedPulling="2026-03-13 08:11:34.869211034 +0000 UTC m=+1954.539990016" lastFinishedPulling="2026-03-13 08:11:35.503796589 +0000 UTC m=+1955.174575571" observedRunningTime="2026-03-13 08:11:36.155392135 +0000 UTC m=+1955.826171117" watchObservedRunningTime="2026-03-13 08:11:36.164160617 +0000 UTC m=+1955.834939609" Mar 13 08:11:44 crc kubenswrapper[4876]: I0313 08:11:44.240213 4876 generic.go:334] "Generic (PLEG): container finished" podID="f87ce9a3-0a04-4b72-bd4f-88a5538c99c1" containerID="f94a02a93179d22563fe0d2062b9bac6c6b6b8f51f84406751d56051543d1734" exitCode=0 Mar 13 08:11:44 crc kubenswrapper[4876]: I0313 08:11:44.240912 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" event={"ID":"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1","Type":"ContainerDied","Data":"f94a02a93179d22563fe0d2062b9bac6c6b6b8f51f84406751d56051543d1734"} Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.678653 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.720373 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-inventory\") pod \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.720448 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz5k4\" (UniqueName: \"kubernetes.io/projected/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-kube-api-access-fz5k4\") pod \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.720586 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-ssh-key-openstack-edpm-ipam\") pod \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\" (UID: \"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1\") " Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.728689 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-kube-api-access-fz5k4" (OuterVolumeSpecName: "kube-api-access-fz5k4") pod "f87ce9a3-0a04-4b72-bd4f-88a5538c99c1" (UID: "f87ce9a3-0a04-4b72-bd4f-88a5538c99c1"). InnerVolumeSpecName "kube-api-access-fz5k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.757109 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f87ce9a3-0a04-4b72-bd4f-88a5538c99c1" (UID: "f87ce9a3-0a04-4b72-bd4f-88a5538c99c1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.759846 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-inventory" (OuterVolumeSpecName: "inventory") pod "f87ce9a3-0a04-4b72-bd4f-88a5538c99c1" (UID: "f87ce9a3-0a04-4b72-bd4f-88a5538c99c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.823084 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.823157 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz5k4\" (UniqueName: \"kubernetes.io/projected/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-kube-api-access-fz5k4\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:45 crc kubenswrapper[4876]: I0313 08:11:45.823175 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f87ce9a3-0a04-4b72-bd4f-88a5538c99c1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.260184 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" event={"ID":"f87ce9a3-0a04-4b72-bd4f-88a5538c99c1","Type":"ContainerDied","Data":"a33688e70e457b7f4c2111925e84071709e17da0ef7744913008d7eca96c2032"} Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.260583 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a33688e70e457b7f4c2111925e84071709e17da0ef7744913008d7eca96c2032" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.260508 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hj2tl" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.339880 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf"] Mar 13 08:11:46 crc kubenswrapper[4876]: E0313 08:11:46.340672 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f87ce9a3-0a04-4b72-bd4f-88a5538c99c1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.340707 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f87ce9a3-0a04-4b72-bd4f-88a5538c99c1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.340975 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f87ce9a3-0a04-4b72-bd4f-88a5538c99c1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.342100 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.345741 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.345762 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.346370 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.346378 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.365783 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf"] Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.434814 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.435318 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.435367 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfc7q\" (UniqueName: \"kubernetes.io/projected/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-kube-api-access-bfc7q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.537397 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfc7q\" (UniqueName: \"kubernetes.io/projected/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-kube-api-access-bfc7q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.537521 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.537603 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.543226 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.550722 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.556195 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfc7q\" (UniqueName: \"kubernetes.io/projected/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-kube-api-access-bfc7q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:46 crc kubenswrapper[4876]: I0313 08:11:46.666945 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:47 crc kubenswrapper[4876]: I0313 08:11:47.199350 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf"] Mar 13 08:11:47 crc kubenswrapper[4876]: I0313 08:11:47.273746 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" event={"ID":"8caf73f8-76f3-4fa6-bfbc-9977576c8d18","Type":"ContainerStarted","Data":"fc8621c1893c642ad77e26b51806710fa1a169d4b84298594eae23a5752d91ae"} Mar 13 08:11:48 crc kubenswrapper[4876]: I0313 08:11:48.282492 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" event={"ID":"8caf73f8-76f3-4fa6-bfbc-9977576c8d18","Type":"ContainerStarted","Data":"516e212ce1dbbf28d4472d6a0644250503e22705cf29a08f2a1b0862d3adeb7c"} Mar 13 08:11:48 crc kubenswrapper[4876]: I0313 08:11:48.305385 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" podStartSLOduration=1.807225734 podStartE2EDuration="2.305361885s" podCreationTimestamp="2026-03-13 08:11:46 +0000 UTC" firstStartedPulling="2026-03-13 08:11:47.209441696 +0000 UTC m=+1966.880220678" lastFinishedPulling="2026-03-13 08:11:47.707577847 +0000 UTC m=+1967.378356829" observedRunningTime="2026-03-13 08:11:48.300671155 +0000 UTC m=+1967.971450137" watchObservedRunningTime="2026-03-13 08:11:48.305361885 +0000 UTC m=+1967.976140867" Mar 13 08:11:58 crc kubenswrapper[4876]: I0313 08:11:58.386795 4876 generic.go:334] "Generic (PLEG): container finished" podID="8caf73f8-76f3-4fa6-bfbc-9977576c8d18" containerID="516e212ce1dbbf28d4472d6a0644250503e22705cf29a08f2a1b0862d3adeb7c" exitCode=0 Mar 13 08:11:58 crc kubenswrapper[4876]: I0313 08:11:58.386911 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" event={"ID":"8caf73f8-76f3-4fa6-bfbc-9977576c8d18","Type":"ContainerDied","Data":"516e212ce1dbbf28d4472d6a0644250503e22705cf29a08f2a1b0862d3adeb7c"} Mar 13 08:11:59 crc kubenswrapper[4876]: I0313 08:11:59.822806 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:11:59 crc kubenswrapper[4876]: I0313 08:11:59.951999 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-ssh-key-openstack-edpm-ipam\") pod \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " Mar 13 08:11:59 crc kubenswrapper[4876]: I0313 08:11:59.952053 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-inventory\") pod \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " Mar 13 08:11:59 crc kubenswrapper[4876]: I0313 08:11:59.952291 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfc7q\" (UniqueName: \"kubernetes.io/projected/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-kube-api-access-bfc7q\") pod \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\" (UID: \"8caf73f8-76f3-4fa6-bfbc-9977576c8d18\") " Mar 13 08:11:59 crc kubenswrapper[4876]: I0313 08:11:59.960500 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-kube-api-access-bfc7q" (OuterVolumeSpecName: "kube-api-access-bfc7q") pod "8caf73f8-76f3-4fa6-bfbc-9977576c8d18" (UID: "8caf73f8-76f3-4fa6-bfbc-9977576c8d18"). InnerVolumeSpecName "kube-api-access-bfc7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:11:59 crc kubenswrapper[4876]: I0313 08:11:59.986466 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-inventory" (OuterVolumeSpecName: "inventory") pod "8caf73f8-76f3-4fa6-bfbc-9977576c8d18" (UID: "8caf73f8-76f3-4fa6-bfbc-9977576c8d18"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:11:59 crc kubenswrapper[4876]: I0313 08:11:59.990559 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8caf73f8-76f3-4fa6-bfbc-9977576c8d18" (UID: "8caf73f8-76f3-4fa6-bfbc-9977576c8d18"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.048448 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-4w2ft"] Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.055075 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.055116 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.055128 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfc7q\" (UniqueName: \"kubernetes.io/projected/8caf73f8-76f3-4fa6-bfbc-9977576c8d18-kube-api-access-bfc7q\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.059494 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-4w2ft"] Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.143543 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556492-x58br"] Mar 13 08:12:00 crc kubenswrapper[4876]: E0313 08:12:00.144892 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8caf73f8-76f3-4fa6-bfbc-9977576c8d18" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.145044 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8caf73f8-76f3-4fa6-bfbc-9977576c8d18" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.145413 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8caf73f8-76f3-4fa6-bfbc-9977576c8d18" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.146431 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556492-x58br" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.150248 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.150586 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.151208 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.153041 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556492-x58br"] Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.258537 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmvqn\" (UniqueName: \"kubernetes.io/projected/272eadd8-ddf0-44c4-8daa-0dcbc371aa6a-kube-api-access-vmvqn\") pod \"auto-csr-approver-29556492-x58br\" (UID: \"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a\") " pod="openshift-infra/auto-csr-approver-29556492-x58br" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.361155 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmvqn\" (UniqueName: \"kubernetes.io/projected/272eadd8-ddf0-44c4-8daa-0dcbc371aa6a-kube-api-access-vmvqn\") pod \"auto-csr-approver-29556492-x58br\" (UID: \"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a\") " pod="openshift-infra/auto-csr-approver-29556492-x58br" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.379742 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmvqn\" (UniqueName: \"kubernetes.io/projected/272eadd8-ddf0-44c4-8daa-0dcbc371aa6a-kube-api-access-vmvqn\") pod \"auto-csr-approver-29556492-x58br\" (UID: \"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a\") " pod="openshift-infra/auto-csr-approver-29556492-x58br" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.417087 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" event={"ID":"8caf73f8-76f3-4fa6-bfbc-9977576c8d18","Type":"ContainerDied","Data":"fc8621c1893c642ad77e26b51806710fa1a169d4b84298594eae23a5752d91ae"} Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.417138 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc8621c1893c642ad77e26b51806710fa1a169d4b84298594eae23a5752d91ae" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.417160 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.470122 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556492-x58br" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.507956 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw"] Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.510401 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.516005 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.525475 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.529912 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.531033 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.531273 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.531456 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.531966 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.532229 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.547028 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw"] Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.669918 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670320 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670407 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670448 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670482 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670509 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670560 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670611 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670645 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670683 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670708 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670724 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670740 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.670767 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzznv\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-kube-api-access-fzznv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772578 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772661 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772690 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772717 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772779 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772843 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772878 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772928 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772961 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.772982 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.773008 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.773045 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzznv\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-kube-api-access-fzznv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.773086 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.773133 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.778142 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.778576 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.779020 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.779749 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.780742 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.781681 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.782044 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.782627 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.783104 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.783860 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.784338 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.784412 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.786790 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.791814 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzznv\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-kube-api-access-fzznv\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xdclw\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.844410 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:00 crc kubenswrapper[4876]: I0313 08:12:00.961151 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556492-x58br"] Mar 13 08:12:01 crc kubenswrapper[4876]: I0313 08:12:01.056625 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41ef81ac-d481-4e47-81d1-c4846012caae" path="/var/lib/kubelet/pods/41ef81ac-d481-4e47-81d1-c4846012caae/volumes" Mar 13 08:12:01 crc kubenswrapper[4876]: I0313 08:12:01.409643 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw"] Mar 13 08:12:01 crc kubenswrapper[4876]: I0313 08:12:01.428157 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" event={"ID":"92d53dff-55bb-4953-96a9-647552c446e0","Type":"ContainerStarted","Data":"c844997f5214f676af4cfd63e5b8421dba3e7a338f0a51e00628a45bbea8de5c"} Mar 13 08:12:01 crc kubenswrapper[4876]: I0313 08:12:01.429478 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556492-x58br" event={"ID":"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a","Type":"ContainerStarted","Data":"228a5954e223b2a81cf3b13bfc1f10b572aea430ba9db06a36cbcd5b7a5c60c0"} Mar 13 08:12:01 crc kubenswrapper[4876]: I0313 08:12:01.989505 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:12:02 crc kubenswrapper[4876]: I0313 08:12:02.439759 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" event={"ID":"92d53dff-55bb-4953-96a9-647552c446e0","Type":"ContainerStarted","Data":"f026f3009b0457f071c240f96ddc64ad2ab115e74758900fc2e11854cd3a5483"} Mar 13 08:12:02 crc kubenswrapper[4876]: I0313 08:12:02.442556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556492-x58br" event={"ID":"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a","Type":"ContainerStarted","Data":"c1ed275ae00f56125e9ec6bdef2d336739470ed3222b67a1e1316622051726ec"} Mar 13 08:12:02 crc kubenswrapper[4876]: I0313 08:12:02.476649 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" podStartSLOduration=1.903301629 podStartE2EDuration="2.476026854s" podCreationTimestamp="2026-03-13 08:12:00 +0000 UTC" firstStartedPulling="2026-03-13 08:12:01.413333992 +0000 UTC m=+1981.084112974" lastFinishedPulling="2026-03-13 08:12:01.986059217 +0000 UTC m=+1981.656838199" observedRunningTime="2026-03-13 08:12:02.468093555 +0000 UTC m=+1982.138872557" watchObservedRunningTime="2026-03-13 08:12:02.476026854 +0000 UTC m=+1982.146805826" Mar 13 08:12:02 crc kubenswrapper[4876]: I0313 08:12:02.497985 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556492-x58br" podStartSLOduration=1.4598422690000001 podStartE2EDuration="2.49795346s" podCreationTimestamp="2026-03-13 08:12:00 +0000 UTC" firstStartedPulling="2026-03-13 08:12:00.97342149 +0000 UTC m=+1980.644200472" lastFinishedPulling="2026-03-13 08:12:02.011532691 +0000 UTC m=+1981.682311663" observedRunningTime="2026-03-13 08:12:02.481909306 +0000 UTC m=+1982.152688288" watchObservedRunningTime="2026-03-13 08:12:02.49795346 +0000 UTC m=+1982.168732452" Mar 13 08:12:03 crc kubenswrapper[4876]: I0313 08:12:03.456168 4876 generic.go:334] "Generic (PLEG): container finished" podID="272eadd8-ddf0-44c4-8daa-0dcbc371aa6a" containerID="c1ed275ae00f56125e9ec6bdef2d336739470ed3222b67a1e1316622051726ec" exitCode=0 Mar 13 08:12:03 crc kubenswrapper[4876]: I0313 08:12:03.456283 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556492-x58br" event={"ID":"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a","Type":"ContainerDied","Data":"c1ed275ae00f56125e9ec6bdef2d336739470ed3222b67a1e1316622051726ec"} Mar 13 08:12:04 crc kubenswrapper[4876]: I0313 08:12:04.828603 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556492-x58br" Mar 13 08:12:04 crc kubenswrapper[4876]: I0313 08:12:04.974765 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmvqn\" (UniqueName: \"kubernetes.io/projected/272eadd8-ddf0-44c4-8daa-0dcbc371aa6a-kube-api-access-vmvqn\") pod \"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a\" (UID: \"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a\") " Mar 13 08:12:04 crc kubenswrapper[4876]: I0313 08:12:04.981428 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/272eadd8-ddf0-44c4-8daa-0dcbc371aa6a-kube-api-access-vmvqn" (OuterVolumeSpecName: "kube-api-access-vmvqn") pod "272eadd8-ddf0-44c4-8daa-0dcbc371aa6a" (UID: "272eadd8-ddf0-44c4-8daa-0dcbc371aa6a"). InnerVolumeSpecName "kube-api-access-vmvqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:12:05 crc kubenswrapper[4876]: I0313 08:12:05.077216 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmvqn\" (UniqueName: \"kubernetes.io/projected/272eadd8-ddf0-44c4-8daa-0dcbc371aa6a-kube-api-access-vmvqn\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:05 crc kubenswrapper[4876]: I0313 08:12:05.475656 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556492-x58br" event={"ID":"272eadd8-ddf0-44c4-8daa-0dcbc371aa6a","Type":"ContainerDied","Data":"228a5954e223b2a81cf3b13bfc1f10b572aea430ba9db06a36cbcd5b7a5c60c0"} Mar 13 08:12:05 crc kubenswrapper[4876]: I0313 08:12:05.476015 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="228a5954e223b2a81cf3b13bfc1f10b572aea430ba9db06a36cbcd5b7a5c60c0" Mar 13 08:12:05 crc kubenswrapper[4876]: I0313 08:12:05.475851 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556492-x58br" Mar 13 08:12:05 crc kubenswrapper[4876]: I0313 08:12:05.896378 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556486-zprrc"] Mar 13 08:12:05 crc kubenswrapper[4876]: I0313 08:12:05.907668 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556486-zprrc"] Mar 13 08:12:07 crc kubenswrapper[4876]: I0313 08:12:07.045562 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="770aa8c1-c4df-4219-bad8-f77761708cb7" path="/var/lib/kubelet/pods/770aa8c1-c4df-4219-bad8-f77761708cb7/volumes" Mar 13 08:12:22 crc kubenswrapper[4876]: I0313 08:12:22.386453 4876 scope.go:117] "RemoveContainer" containerID="f56c3f230c2f11f80c75e9b403149d9aefb10f7aaefb8d78e497bbb88bb623a6" Mar 13 08:12:22 crc kubenswrapper[4876]: I0313 08:12:22.412683 4876 scope.go:117] "RemoveContainer" containerID="fa619c02492710df3118fa5e92ce5fed9bbddf8b3dc470b1fd01fc684d4f2942" Mar 13 08:12:22 crc kubenswrapper[4876]: I0313 08:12:22.440055 4876 scope.go:117] "RemoveContainer" containerID="31be08a512386fe4f1a39fa7ed417b42ad05e75f220adb51f58fb1cf7ae50103" Mar 13 08:12:22 crc kubenswrapper[4876]: I0313 08:12:22.570834 4876 scope.go:117] "RemoveContainer" containerID="056f5011fdbb1d4f4c1dc934977d34db0e3d9741d5b5853de942e1a8a0b6e872" Mar 13 08:12:22 crc kubenswrapper[4876]: I0313 08:12:22.630923 4876 scope.go:117] "RemoveContainer" containerID="9c805176a408f3ddb2df890a616db47edb2d137dc31bb29a5baacfe27370b7ab" Mar 13 08:12:37 crc kubenswrapper[4876]: I0313 08:12:37.850664 4876 generic.go:334] "Generic (PLEG): container finished" podID="92d53dff-55bb-4953-96a9-647552c446e0" containerID="f026f3009b0457f071c240f96ddc64ad2ab115e74758900fc2e11854cd3a5483" exitCode=0 Mar 13 08:12:37 crc kubenswrapper[4876]: I0313 08:12:37.850788 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" event={"ID":"92d53dff-55bb-4953-96a9-647552c446e0","Type":"ContainerDied","Data":"f026f3009b0457f071c240f96ddc64ad2ab115e74758900fc2e11854cd3a5483"} Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.352055 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506641 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ovn-combined-ca-bundle\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506703 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzznv\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-kube-api-access-fzznv\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506732 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-libvirt-combined-ca-bundle\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506768 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-ovn-default-certs-0\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506842 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506877 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-repo-setup-combined-ca-bundle\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506897 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-inventory\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506935 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-bootstrap-combined-ca-bundle\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.506969 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-telemetry-combined-ca-bundle\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.507038 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ssh-key-openstack-edpm-ipam\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.507071 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-nova-combined-ca-bundle\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.507096 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.507164 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-neutron-metadata-combined-ca-bundle\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.507210 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"92d53dff-55bb-4953-96a9-647552c446e0\" (UID: \"92d53dff-55bb-4953-96a9-647552c446e0\") " Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.514529 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-kube-api-access-fzznv" (OuterVolumeSpecName: "kube-api-access-fzznv") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "kube-api-access-fzznv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.515141 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.514588 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.515040 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.515791 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.519000 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.521500 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.521608 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.522302 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.522514 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.522962 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.527672 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.547454 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.554458 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-inventory" (OuterVolumeSpecName: "inventory") pod "92d53dff-55bb-4953-96a9-647552c446e0" (UID: "92d53dff-55bb-4953-96a9-647552c446e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.610568 4876 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611132 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611224 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611333 4876 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611439 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzznv\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-kube-api-access-fzznv\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611559 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611676 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611777 4876 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611874 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.611967 4876 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.612049 4876 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.612329 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.612443 4876 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53dff-55bb-4953-96a9-647552c446e0-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.612531 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d53dff-55bb-4953-96a9-647552c446e0-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.874569 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" event={"ID":"92d53dff-55bb-4953-96a9-647552c446e0","Type":"ContainerDied","Data":"c844997f5214f676af4cfd63e5b8421dba3e7a338f0a51e00628a45bbea8de5c"} Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.874704 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c844997f5214f676af4cfd63e5b8421dba3e7a338f0a51e00628a45bbea8de5c" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.874638 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xdclw" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.987141 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj"] Mar 13 08:12:39 crc kubenswrapper[4876]: E0313 08:12:39.987699 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="272eadd8-ddf0-44c4-8daa-0dcbc371aa6a" containerName="oc" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.987721 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="272eadd8-ddf0-44c4-8daa-0dcbc371aa6a" containerName="oc" Mar 13 08:12:39 crc kubenswrapper[4876]: E0313 08:12:39.987752 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d53dff-55bb-4953-96a9-647552c446e0" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.987762 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d53dff-55bb-4953-96a9-647552c446e0" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.987957 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="272eadd8-ddf0-44c4-8daa-0dcbc371aa6a" containerName="oc" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.987970 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d53dff-55bb-4953-96a9-647552c446e0" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.988767 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.993706 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.993998 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.994164 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.994386 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:12:39 crc kubenswrapper[4876]: I0313 08:12:39.994555 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.005222 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj"] Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.030414 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.030630 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/848b0256-b349-49c3-a007-20c0da303ad8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.030791 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.031063 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.031210 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxbnt\" (UniqueName: \"kubernetes.io/projected/848b0256-b349-49c3-a007-20c0da303ad8-kube-api-access-dxbnt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.133580 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.133786 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.133839 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxbnt\" (UniqueName: \"kubernetes.io/projected/848b0256-b349-49c3-a007-20c0da303ad8-kube-api-access-dxbnt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.133927 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.134041 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/848b0256-b349-49c3-a007-20c0da303ad8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.135269 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/848b0256-b349-49c3-a007-20c0da303ad8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.140034 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.140473 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.144919 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.153116 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxbnt\" (UniqueName: \"kubernetes.io/projected/848b0256-b349-49c3-a007-20c0da303ad8-kube-api-access-dxbnt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hrnbj\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.314726 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.858411 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj"] Mar 13 08:12:40 crc kubenswrapper[4876]: I0313 08:12:40.884081 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" event={"ID":"848b0256-b349-49c3-a007-20c0da303ad8","Type":"ContainerStarted","Data":"c0481eb74708a2a85fea77f8854cac3caeffb23174aae834106925e965308424"} Mar 13 08:12:41 crc kubenswrapper[4876]: I0313 08:12:41.894899 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" event={"ID":"848b0256-b349-49c3-a007-20c0da303ad8","Type":"ContainerStarted","Data":"7dc2411e3f8fe598542f65887de8ecded5b081c72002ae530cb0dcbcb19e032d"} Mar 13 08:12:41 crc kubenswrapper[4876]: I0313 08:12:41.921268 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" podStartSLOduration=2.511359707 podStartE2EDuration="2.921221438s" podCreationTimestamp="2026-03-13 08:12:39 +0000 UTC" firstStartedPulling="2026-03-13 08:12:40.857294204 +0000 UTC m=+2020.528073186" lastFinishedPulling="2026-03-13 08:12:41.267155915 +0000 UTC m=+2020.937934917" observedRunningTime="2026-03-13 08:12:41.911526031 +0000 UTC m=+2021.582305023" watchObservedRunningTime="2026-03-13 08:12:41.921221438 +0000 UTC m=+2021.592000430" Mar 13 08:13:24 crc kubenswrapper[4876]: I0313 08:13:24.608092 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:13:24 crc kubenswrapper[4876]: I0313 08:13:24.608819 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:13:43 crc kubenswrapper[4876]: I0313 08:13:43.525518 4876 generic.go:334] "Generic (PLEG): container finished" podID="848b0256-b349-49c3-a007-20c0da303ad8" containerID="7dc2411e3f8fe598542f65887de8ecded5b081c72002ae530cb0dcbcb19e032d" exitCode=0 Mar 13 08:13:43 crc kubenswrapper[4876]: I0313 08:13:43.525593 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" event={"ID":"848b0256-b349-49c3-a007-20c0da303ad8","Type":"ContainerDied","Data":"7dc2411e3f8fe598542f65887de8ecded5b081c72002ae530cb0dcbcb19e032d"} Mar 13 08:13:44 crc kubenswrapper[4876]: I0313 08:13:44.944969 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.096918 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-inventory\") pod \"848b0256-b349-49c3-a007-20c0da303ad8\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.097654 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxbnt\" (UniqueName: \"kubernetes.io/projected/848b0256-b349-49c3-a007-20c0da303ad8-kube-api-access-dxbnt\") pod \"848b0256-b349-49c3-a007-20c0da303ad8\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.097748 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ssh-key-openstack-edpm-ipam\") pod \"848b0256-b349-49c3-a007-20c0da303ad8\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.097855 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/848b0256-b349-49c3-a007-20c0da303ad8-ovncontroller-config-0\") pod \"848b0256-b349-49c3-a007-20c0da303ad8\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.098060 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ovn-combined-ca-bundle\") pod \"848b0256-b349-49c3-a007-20c0da303ad8\" (UID: \"848b0256-b349-49c3-a007-20c0da303ad8\") " Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.104939 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/848b0256-b349-49c3-a007-20c0da303ad8-kube-api-access-dxbnt" (OuterVolumeSpecName: "kube-api-access-dxbnt") pod "848b0256-b349-49c3-a007-20c0da303ad8" (UID: "848b0256-b349-49c3-a007-20c0da303ad8"). InnerVolumeSpecName "kube-api-access-dxbnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.105078 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "848b0256-b349-49c3-a007-20c0da303ad8" (UID: "848b0256-b349-49c3-a007-20c0da303ad8"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.131534 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "848b0256-b349-49c3-a007-20c0da303ad8" (UID: "848b0256-b349-49c3-a007-20c0da303ad8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.137429 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-inventory" (OuterVolumeSpecName: "inventory") pod "848b0256-b349-49c3-a007-20c0da303ad8" (UID: "848b0256-b349-49c3-a007-20c0da303ad8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.137613 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/848b0256-b349-49c3-a007-20c0da303ad8-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "848b0256-b349-49c3-a007-20c0da303ad8" (UID: "848b0256-b349-49c3-a007-20c0da303ad8"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.200812 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.200844 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.200854 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxbnt\" (UniqueName: \"kubernetes.io/projected/848b0256-b349-49c3-a007-20c0da303ad8-kube-api-access-dxbnt\") on node \"crc\" DevicePath \"\"" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.200865 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848b0256-b349-49c3-a007-20c0da303ad8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.200874 4876 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/848b0256-b349-49c3-a007-20c0da303ad8-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.544078 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" event={"ID":"848b0256-b349-49c3-a007-20c0da303ad8","Type":"ContainerDied","Data":"c0481eb74708a2a85fea77f8854cac3caeffb23174aae834106925e965308424"} Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.544138 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0481eb74708a2a85fea77f8854cac3caeffb23174aae834106925e965308424" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.544202 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hrnbj" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.662144 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg"] Mar 13 08:13:45 crc kubenswrapper[4876]: E0313 08:13:45.662576 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="848b0256-b349-49c3-a007-20c0da303ad8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.662596 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="848b0256-b349-49c3-a007-20c0da303ad8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.662760 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="848b0256-b349-49c3-a007-20c0da303ad8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.663580 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.665720 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.665969 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.666137 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.666361 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.666518 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.673190 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg"] Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.673868 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.811295 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrgr8\" (UniqueName: \"kubernetes.io/projected/97951031-bd23-462e-83a8-be7068e1dfc8-kube-api-access-zrgr8\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.811418 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.811449 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.811470 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.811626 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.811683 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.912847 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.912903 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.912957 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.912977 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.913025 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrgr8\" (UniqueName: \"kubernetes.io/projected/97951031-bd23-462e-83a8-be7068e1dfc8-kube-api-access-zrgr8\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.913102 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.919386 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.919485 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.922231 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.922588 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.923489 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.931539 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrgr8\" (UniqueName: \"kubernetes.io/projected/97951031-bd23-462e-83a8-be7068e1dfc8-kube-api-access-zrgr8\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:45 crc kubenswrapper[4876]: I0313 08:13:45.983543 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:13:46 crc kubenswrapper[4876]: I0313 08:13:46.538787 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg"] Mar 13 08:13:46 crc kubenswrapper[4876]: I0313 08:13:46.548950 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:13:46 crc kubenswrapper[4876]: I0313 08:13:46.560981 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" event={"ID":"97951031-bd23-462e-83a8-be7068e1dfc8","Type":"ContainerStarted","Data":"d78763dd3bb8fef845425617d067b3c34c3a010f521cca5252094f6e35a016d7"} Mar 13 08:13:47 crc kubenswrapper[4876]: I0313 08:13:47.571965 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" event={"ID":"97951031-bd23-462e-83a8-be7068e1dfc8","Type":"ContainerStarted","Data":"c120fc114397276546b4b7c18e30eefdc214c70f5cc5be641b6c2a98832d20b0"} Mar 13 08:13:47 crc kubenswrapper[4876]: I0313 08:13:47.594586 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" podStartSLOduration=2.11108772 podStartE2EDuration="2.594558667s" podCreationTimestamp="2026-03-13 08:13:45 +0000 UTC" firstStartedPulling="2026-03-13 08:13:46.548572397 +0000 UTC m=+2086.219351379" lastFinishedPulling="2026-03-13 08:13:47.032043344 +0000 UTC m=+2086.702822326" observedRunningTime="2026-03-13 08:13:47.588651323 +0000 UTC m=+2087.259430305" watchObservedRunningTime="2026-03-13 08:13:47.594558667 +0000 UTC m=+2087.265337649" Mar 13 08:13:54 crc kubenswrapper[4876]: I0313 08:13:54.607373 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:13:54 crc kubenswrapper[4876]: I0313 08:13:54.608059 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.161210 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556494-69d6c"] Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.164603 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556494-69d6c" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.167390 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.168209 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.170266 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.176881 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556494-69d6c"] Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.295852 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj5j7\" (UniqueName: \"kubernetes.io/projected/d9d16336-44b2-4142-a4e7-d0770f7a96cd-kube-api-access-pj5j7\") pod \"auto-csr-approver-29556494-69d6c\" (UID: \"d9d16336-44b2-4142-a4e7-d0770f7a96cd\") " pod="openshift-infra/auto-csr-approver-29556494-69d6c" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.397737 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj5j7\" (UniqueName: \"kubernetes.io/projected/d9d16336-44b2-4142-a4e7-d0770f7a96cd-kube-api-access-pj5j7\") pod \"auto-csr-approver-29556494-69d6c\" (UID: \"d9d16336-44b2-4142-a4e7-d0770f7a96cd\") " pod="openshift-infra/auto-csr-approver-29556494-69d6c" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.424368 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj5j7\" (UniqueName: \"kubernetes.io/projected/d9d16336-44b2-4142-a4e7-d0770f7a96cd-kube-api-access-pj5j7\") pod \"auto-csr-approver-29556494-69d6c\" (UID: \"d9d16336-44b2-4142-a4e7-d0770f7a96cd\") " pod="openshift-infra/auto-csr-approver-29556494-69d6c" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.492481 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556494-69d6c" Mar 13 08:14:00 crc kubenswrapper[4876]: I0313 08:14:00.950938 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556494-69d6c"] Mar 13 08:14:01 crc kubenswrapper[4876]: I0313 08:14:01.731164 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556494-69d6c" event={"ID":"d9d16336-44b2-4142-a4e7-d0770f7a96cd","Type":"ContainerStarted","Data":"119a16934c3487b0591987f8d5905af6b6f699f567a25db2a05c0817099e3571"} Mar 13 08:14:02 crc kubenswrapper[4876]: I0313 08:14:02.741392 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556494-69d6c" event={"ID":"d9d16336-44b2-4142-a4e7-d0770f7a96cd","Type":"ContainerStarted","Data":"d42bb885c59a8700da179f0ae109e5eaa5f35cde19b9357d32fae60d0c9302f5"} Mar 13 08:14:02 crc kubenswrapper[4876]: I0313 08:14:02.760740 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556494-69d6c" podStartSLOduration=1.354833548 podStartE2EDuration="2.760713578s" podCreationTimestamp="2026-03-13 08:14:00 +0000 UTC" firstStartedPulling="2026-03-13 08:14:00.960137646 +0000 UTC m=+2100.630916628" lastFinishedPulling="2026-03-13 08:14:02.366017676 +0000 UTC m=+2102.036796658" observedRunningTime="2026-03-13 08:14:02.758003774 +0000 UTC m=+2102.428782756" watchObservedRunningTime="2026-03-13 08:14:02.760713578 +0000 UTC m=+2102.431492570" Mar 13 08:14:03 crc kubenswrapper[4876]: I0313 08:14:03.772887 4876 generic.go:334] "Generic (PLEG): container finished" podID="d9d16336-44b2-4142-a4e7-d0770f7a96cd" containerID="d42bb885c59a8700da179f0ae109e5eaa5f35cde19b9357d32fae60d0c9302f5" exitCode=0 Mar 13 08:14:03 crc kubenswrapper[4876]: I0313 08:14:03.772968 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556494-69d6c" event={"ID":"d9d16336-44b2-4142-a4e7-d0770f7a96cd","Type":"ContainerDied","Data":"d42bb885c59a8700da179f0ae109e5eaa5f35cde19b9357d32fae60d0c9302f5"} Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.091175 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556494-69d6c" Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.194897 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj5j7\" (UniqueName: \"kubernetes.io/projected/d9d16336-44b2-4142-a4e7-d0770f7a96cd-kube-api-access-pj5j7\") pod \"d9d16336-44b2-4142-a4e7-d0770f7a96cd\" (UID: \"d9d16336-44b2-4142-a4e7-d0770f7a96cd\") " Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.200986 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d16336-44b2-4142-a4e7-d0770f7a96cd-kube-api-access-pj5j7" (OuterVolumeSpecName: "kube-api-access-pj5j7") pod "d9d16336-44b2-4142-a4e7-d0770f7a96cd" (UID: "d9d16336-44b2-4142-a4e7-d0770f7a96cd"). InnerVolumeSpecName "kube-api-access-pj5j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.297465 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj5j7\" (UniqueName: \"kubernetes.io/projected/d9d16336-44b2-4142-a4e7-d0770f7a96cd-kube-api-access-pj5j7\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.794334 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556494-69d6c" event={"ID":"d9d16336-44b2-4142-a4e7-d0770f7a96cd","Type":"ContainerDied","Data":"119a16934c3487b0591987f8d5905af6b6f699f567a25db2a05c0817099e3571"} Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.794385 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="119a16934c3487b0591987f8d5905af6b6f699f567a25db2a05c0817099e3571" Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.794470 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556494-69d6c" Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.850985 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556488-mwrfn"] Mar 13 08:14:05 crc kubenswrapper[4876]: I0313 08:14:05.859543 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556488-mwrfn"] Mar 13 08:14:07 crc kubenswrapper[4876]: I0313 08:14:07.046350 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5452b7b8-0dbd-46f7-91d4-2321bcb62f5f" path="/var/lib/kubelet/pods/5452b7b8-0dbd-46f7-91d4-2321bcb62f5f/volumes" Mar 13 08:14:22 crc kubenswrapper[4876]: I0313 08:14:22.737913 4876 scope.go:117] "RemoveContainer" containerID="8c2d6cd129b5cf087e15db2b85a5464636c60bfb51c76696d6f692964d49e6aa" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.003531 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vlngd"] Mar 13 08:14:23 crc kubenswrapper[4876]: E0313 08:14:23.004323 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d16336-44b2-4142-a4e7-d0770f7a96cd" containerName="oc" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.004342 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d16336-44b2-4142-a4e7-d0770f7a96cd" containerName="oc" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.004534 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d16336-44b2-4142-a4e7-d0770f7a96cd" containerName="oc" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.005867 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vlngd"] Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.005954 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.072747 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwvc4\" (UniqueName: \"kubernetes.io/projected/4d209011-013d-4550-9188-e49679afac64-kube-api-access-nwvc4\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.072817 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-utilities\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.072925 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-catalog-content\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.174170 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-catalog-content\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.174765 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-catalog-content\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.174903 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwvc4\" (UniqueName: \"kubernetes.io/projected/4d209011-013d-4550-9188-e49679afac64-kube-api-access-nwvc4\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.175311 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-utilities\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.175653 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-utilities\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.198835 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwvc4\" (UniqueName: \"kubernetes.io/projected/4d209011-013d-4550-9188-e49679afac64-kube-api-access-nwvc4\") pod \"redhat-operators-vlngd\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.371285 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.876431 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vlngd"] Mar 13 08:14:23 crc kubenswrapper[4876]: I0313 08:14:23.980619 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlngd" event={"ID":"4d209011-013d-4550-9188-e49679afac64","Type":"ContainerStarted","Data":"5330e2e667e70d2170cc94b1578673d8bf9f6cccc365e3438d46d0099824b3c5"} Mar 13 08:14:24 crc kubenswrapper[4876]: I0313 08:14:24.607702 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:14:24 crc kubenswrapper[4876]: I0313 08:14:24.608091 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:14:24 crc kubenswrapper[4876]: I0313 08:14:24.608150 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:14:24 crc kubenswrapper[4876]: I0313 08:14:24.608758 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0c73df6ed9f0626ecd967ee4e119fdb9bc95174ba5b98bf45405cb2d0c94943"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:14:24 crc kubenswrapper[4876]: I0313 08:14:24.608823 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://d0c73df6ed9f0626ecd967ee4e119fdb9bc95174ba5b98bf45405cb2d0c94943" gracePeriod=600 Mar 13 08:14:24 crc kubenswrapper[4876]: I0313 08:14:24.995801 4876 generic.go:334] "Generic (PLEG): container finished" podID="4d209011-013d-4550-9188-e49679afac64" containerID="c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457" exitCode=0 Mar 13 08:14:24 crc kubenswrapper[4876]: I0313 08:14:24.995907 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlngd" event={"ID":"4d209011-013d-4550-9188-e49679afac64","Type":"ContainerDied","Data":"c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457"} Mar 13 08:14:25 crc kubenswrapper[4876]: I0313 08:14:25.001064 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="d0c73df6ed9f0626ecd967ee4e119fdb9bc95174ba5b98bf45405cb2d0c94943" exitCode=0 Mar 13 08:14:25 crc kubenswrapper[4876]: I0313 08:14:25.001129 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"d0c73df6ed9f0626ecd967ee4e119fdb9bc95174ba5b98bf45405cb2d0c94943"} Mar 13 08:14:25 crc kubenswrapper[4876]: I0313 08:14:25.001178 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66"} Mar 13 08:14:25 crc kubenswrapper[4876]: I0313 08:14:25.001224 4876 scope.go:117] "RemoveContainer" containerID="880ac14da20b999410f87cf1764e08a17bd154e6d6b78407a43fabc808b48ca2" Mar 13 08:14:26 crc kubenswrapper[4876]: I0313 08:14:26.015535 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlngd" event={"ID":"4d209011-013d-4550-9188-e49679afac64","Type":"ContainerStarted","Data":"33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532"} Mar 13 08:14:27 crc kubenswrapper[4876]: I0313 08:14:27.030263 4876 generic.go:334] "Generic (PLEG): container finished" podID="4d209011-013d-4550-9188-e49679afac64" containerID="33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532" exitCode=0 Mar 13 08:14:27 crc kubenswrapper[4876]: I0313 08:14:27.030376 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlngd" event={"ID":"4d209011-013d-4550-9188-e49679afac64","Type":"ContainerDied","Data":"33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532"} Mar 13 08:14:28 crc kubenswrapper[4876]: I0313 08:14:28.060354 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlngd" event={"ID":"4d209011-013d-4550-9188-e49679afac64","Type":"ContainerStarted","Data":"96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab"} Mar 13 08:14:28 crc kubenswrapper[4876]: I0313 08:14:28.086720 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vlngd" podStartSLOduration=3.3509263000000002 podStartE2EDuration="6.086698508s" podCreationTimestamp="2026-03-13 08:14:22 +0000 UTC" firstStartedPulling="2026-03-13 08:14:24.999317738 +0000 UTC m=+2124.670096760" lastFinishedPulling="2026-03-13 08:14:27.735089986 +0000 UTC m=+2127.405868968" observedRunningTime="2026-03-13 08:14:28.077100992 +0000 UTC m=+2127.747879974" watchObservedRunningTime="2026-03-13 08:14:28.086698508 +0000 UTC m=+2127.757477500" Mar 13 08:14:33 crc kubenswrapper[4876]: I0313 08:14:33.372432 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:33 crc kubenswrapper[4876]: I0313 08:14:33.373067 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:34 crc kubenswrapper[4876]: I0313 08:14:34.416061 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vlngd" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="registry-server" probeResult="failure" output=< Mar 13 08:14:34 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:14:34 crc kubenswrapper[4876]: > Mar 13 08:14:35 crc kubenswrapper[4876]: I0313 08:14:35.124656 4876 generic.go:334] "Generic (PLEG): container finished" podID="97951031-bd23-462e-83a8-be7068e1dfc8" containerID="c120fc114397276546b4b7c18e30eefdc214c70f5cc5be641b6c2a98832d20b0" exitCode=0 Mar 13 08:14:35 crc kubenswrapper[4876]: I0313 08:14:35.124708 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" event={"ID":"97951031-bd23-462e-83a8-be7068e1dfc8","Type":"ContainerDied","Data":"c120fc114397276546b4b7c18e30eefdc214c70f5cc5be641b6c2a98832d20b0"} Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.560662 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.631671 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"97951031-bd23-462e-83a8-be7068e1dfc8\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.631715 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrgr8\" (UniqueName: \"kubernetes.io/projected/97951031-bd23-462e-83a8-be7068e1dfc8-kube-api-access-zrgr8\") pod \"97951031-bd23-462e-83a8-be7068e1dfc8\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.631876 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-inventory\") pod \"97951031-bd23-462e-83a8-be7068e1dfc8\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.631972 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-metadata-combined-ca-bundle\") pod \"97951031-bd23-462e-83a8-be7068e1dfc8\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.632018 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-nova-metadata-neutron-config-0\") pod \"97951031-bd23-462e-83a8-be7068e1dfc8\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.632093 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-ssh-key-openstack-edpm-ipam\") pod \"97951031-bd23-462e-83a8-be7068e1dfc8\" (UID: \"97951031-bd23-462e-83a8-be7068e1dfc8\") " Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.637580 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97951031-bd23-462e-83a8-be7068e1dfc8-kube-api-access-zrgr8" (OuterVolumeSpecName: "kube-api-access-zrgr8") pod "97951031-bd23-462e-83a8-be7068e1dfc8" (UID: "97951031-bd23-462e-83a8-be7068e1dfc8"). InnerVolumeSpecName "kube-api-access-zrgr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.638074 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "97951031-bd23-462e-83a8-be7068e1dfc8" (UID: "97951031-bd23-462e-83a8-be7068e1dfc8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.660716 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "97951031-bd23-462e-83a8-be7068e1dfc8" (UID: "97951031-bd23-462e-83a8-be7068e1dfc8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.661225 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "97951031-bd23-462e-83a8-be7068e1dfc8" (UID: "97951031-bd23-462e-83a8-be7068e1dfc8"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.666409 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "97951031-bd23-462e-83a8-be7068e1dfc8" (UID: "97951031-bd23-462e-83a8-be7068e1dfc8"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.666433 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-inventory" (OuterVolumeSpecName: "inventory") pod "97951031-bd23-462e-83a8-be7068e1dfc8" (UID: "97951031-bd23-462e-83a8-be7068e1dfc8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.735056 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.735121 4876 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.735141 4876 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.735153 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.735166 4876 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/97951031-bd23-462e-83a8-be7068e1dfc8-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:36 crc kubenswrapper[4876]: I0313 08:14:36.735178 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrgr8\" (UniqueName: \"kubernetes.io/projected/97951031-bd23-462e-83a8-be7068e1dfc8-kube-api-access-zrgr8\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.145200 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" event={"ID":"97951031-bd23-462e-83a8-be7068e1dfc8","Type":"ContainerDied","Data":"d78763dd3bb8fef845425617d067b3c34c3a010f521cca5252094f6e35a016d7"} Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.145272 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d78763dd3bb8fef845425617d067b3c34c3a010f521cca5252094f6e35a016d7" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.145244 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.235762 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k"] Mar 13 08:14:37 crc kubenswrapper[4876]: E0313 08:14:37.236270 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97951031-bd23-462e-83a8-be7068e1dfc8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.236294 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="97951031-bd23-462e-83a8-be7068e1dfc8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.236563 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="97951031-bd23-462e-83a8-be7068e1dfc8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.237170 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.238903 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.239571 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.239994 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.240334 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.240712 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.254364 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k"] Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.344122 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.344199 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.344278 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7pnj\" (UniqueName: \"kubernetes.io/projected/240f0880-695b-44b4-ae52-2beacd34f2d0-kube-api-access-p7pnj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.344453 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.344598 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.446689 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.446779 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7pnj\" (UniqueName: \"kubernetes.io/projected/240f0880-695b-44b4-ae52-2beacd34f2d0-kube-api-access-p7pnj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.446823 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.446855 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.446943 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.451843 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.454836 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.455116 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.456818 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.464167 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7pnj\" (UniqueName: \"kubernetes.io/projected/240f0880-695b-44b4-ae52-2beacd34f2d0-kube-api-access-p7pnj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:37 crc kubenswrapper[4876]: I0313 08:14:37.555280 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:14:38 crc kubenswrapper[4876]: I0313 08:14:38.116276 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k"] Mar 13 08:14:38 crc kubenswrapper[4876]: I0313 08:14:38.156360 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" event={"ID":"240f0880-695b-44b4-ae52-2beacd34f2d0","Type":"ContainerStarted","Data":"3cec191ec1de73416f5f6ec968fdb3a267722005b7b91f3eea9a8a3ac761480f"} Mar 13 08:14:39 crc kubenswrapper[4876]: I0313 08:14:39.164849 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" event={"ID":"240f0880-695b-44b4-ae52-2beacd34f2d0","Type":"ContainerStarted","Data":"6a01fecc812f6430ee52ed38fa49922eae8e8e2061980b991f2cb03a4ba9c28e"} Mar 13 08:14:39 crc kubenswrapper[4876]: I0313 08:14:39.182132 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" podStartSLOduration=1.564845822 podStartE2EDuration="2.182112412s" podCreationTimestamp="2026-03-13 08:14:37 +0000 UTC" firstStartedPulling="2026-03-13 08:14:38.113766852 +0000 UTC m=+2137.784545834" lastFinishedPulling="2026-03-13 08:14:38.731033402 +0000 UTC m=+2138.401812424" observedRunningTime="2026-03-13 08:14:39.180148129 +0000 UTC m=+2138.850927121" watchObservedRunningTime="2026-03-13 08:14:39.182112412 +0000 UTC m=+2138.852891394" Mar 13 08:14:43 crc kubenswrapper[4876]: I0313 08:14:43.415116 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:43 crc kubenswrapper[4876]: I0313 08:14:43.459535 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:43 crc kubenswrapper[4876]: I0313 08:14:43.654562 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vlngd"] Mar 13 08:14:45 crc kubenswrapper[4876]: I0313 08:14:45.217661 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vlngd" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="registry-server" containerID="cri-o://96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab" gracePeriod=2 Mar 13 08:14:45 crc kubenswrapper[4876]: I0313 08:14:45.772301 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:45 crc kubenswrapper[4876]: I0313 08:14:45.917674 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-utilities\") pod \"4d209011-013d-4550-9188-e49679afac64\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " Mar 13 08:14:45 crc kubenswrapper[4876]: I0313 08:14:45.917883 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-catalog-content\") pod \"4d209011-013d-4550-9188-e49679afac64\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " Mar 13 08:14:45 crc kubenswrapper[4876]: I0313 08:14:45.918018 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwvc4\" (UniqueName: \"kubernetes.io/projected/4d209011-013d-4550-9188-e49679afac64-kube-api-access-nwvc4\") pod \"4d209011-013d-4550-9188-e49679afac64\" (UID: \"4d209011-013d-4550-9188-e49679afac64\") " Mar 13 08:14:45 crc kubenswrapper[4876]: I0313 08:14:45.918671 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-utilities" (OuterVolumeSpecName: "utilities") pod "4d209011-013d-4550-9188-e49679afac64" (UID: "4d209011-013d-4550-9188-e49679afac64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:14:45 crc kubenswrapper[4876]: I0313 08:14:45.923858 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d209011-013d-4550-9188-e49679afac64-kube-api-access-nwvc4" (OuterVolumeSpecName: "kube-api-access-nwvc4") pod "4d209011-013d-4550-9188-e49679afac64" (UID: "4d209011-013d-4550-9188-e49679afac64"). InnerVolumeSpecName "kube-api-access-nwvc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.020644 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwvc4\" (UniqueName: \"kubernetes.io/projected/4d209011-013d-4550-9188-e49679afac64-kube-api-access-nwvc4\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.020682 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.060473 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d209011-013d-4550-9188-e49679afac64" (UID: "4d209011-013d-4550-9188-e49679afac64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.122093 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d209011-013d-4550-9188-e49679afac64-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.228269 4876 generic.go:334] "Generic (PLEG): container finished" podID="4d209011-013d-4550-9188-e49679afac64" containerID="96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab" exitCode=0 Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.228310 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlngd" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.228325 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlngd" event={"ID":"4d209011-013d-4550-9188-e49679afac64","Type":"ContainerDied","Data":"96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab"} Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.228368 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlngd" event={"ID":"4d209011-013d-4550-9188-e49679afac64","Type":"ContainerDied","Data":"5330e2e667e70d2170cc94b1578673d8bf9f6cccc365e3438d46d0099824b3c5"} Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.228389 4876 scope.go:117] "RemoveContainer" containerID="96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.260080 4876 scope.go:117] "RemoveContainer" containerID="33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.262636 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vlngd"] Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.271141 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vlngd"] Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.296215 4876 scope.go:117] "RemoveContainer" containerID="c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.331418 4876 scope.go:117] "RemoveContainer" containerID="96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab" Mar 13 08:14:46 crc kubenswrapper[4876]: E0313 08:14:46.333986 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab\": container with ID starting with 96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab not found: ID does not exist" containerID="96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.334027 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab"} err="failed to get container status \"96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab\": rpc error: code = NotFound desc = could not find container \"96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab\": container with ID starting with 96e34ced19dc0b23d1138421612c05a9b3f5af74f6e4c60ba1e47f7a3fb162ab not found: ID does not exist" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.334054 4876 scope.go:117] "RemoveContainer" containerID="33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532" Mar 13 08:14:46 crc kubenswrapper[4876]: E0313 08:14:46.334471 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532\": container with ID starting with 33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532 not found: ID does not exist" containerID="33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.334504 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532"} err="failed to get container status \"33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532\": rpc error: code = NotFound desc = could not find container \"33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532\": container with ID starting with 33398e4c80576f449f255ce052c1c6afbc9963d57cd77b961dd31c88e2e87532 not found: ID does not exist" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.334526 4876 scope.go:117] "RemoveContainer" containerID="c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457" Mar 13 08:14:46 crc kubenswrapper[4876]: E0313 08:14:46.334891 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457\": container with ID starting with c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457 not found: ID does not exist" containerID="c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457" Mar 13 08:14:46 crc kubenswrapper[4876]: I0313 08:14:46.334914 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457"} err="failed to get container status \"c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457\": rpc error: code = NotFound desc = could not find container \"c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457\": container with ID starting with c1cf57611eea365869cdee0617b9458871b75ded22a2c8168f6290e57819e457 not found: ID does not exist" Mar 13 08:14:47 crc kubenswrapper[4876]: I0313 08:14:47.049006 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d209011-013d-4550-9188-e49679afac64" path="/var/lib/kubelet/pods/4d209011-013d-4550-9188-e49679afac64/volumes" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.157402 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8"] Mar 13 08:15:00 crc kubenswrapper[4876]: E0313 08:15:00.158406 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="registry-server" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.158422 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="registry-server" Mar 13 08:15:00 crc kubenswrapper[4876]: E0313 08:15:00.158443 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="extract-content" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.158452 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="extract-content" Mar 13 08:15:00 crc kubenswrapper[4876]: E0313 08:15:00.158479 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="extract-utilities" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.158488 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="extract-utilities" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.158718 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d209011-013d-4550-9188-e49679afac64" containerName="registry-server" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.159456 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.162748 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.163312 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.167435 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8"] Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.195552 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr5r8\" (UniqueName: \"kubernetes.io/projected/a44c13b9-7f84-4443-ac9f-52687a97990f-kube-api-access-lr5r8\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.195651 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a44c13b9-7f84-4443-ac9f-52687a97990f-config-volume\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.195753 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a44c13b9-7f84-4443-ac9f-52687a97990f-secret-volume\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.296686 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a44c13b9-7f84-4443-ac9f-52687a97990f-secret-volume\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.296849 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr5r8\" (UniqueName: \"kubernetes.io/projected/a44c13b9-7f84-4443-ac9f-52687a97990f-kube-api-access-lr5r8\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.296909 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a44c13b9-7f84-4443-ac9f-52687a97990f-config-volume\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.297896 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a44c13b9-7f84-4443-ac9f-52687a97990f-config-volume\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.305566 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a44c13b9-7f84-4443-ac9f-52687a97990f-secret-volume\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.323516 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr5r8\" (UniqueName: \"kubernetes.io/projected/a44c13b9-7f84-4443-ac9f-52687a97990f-kube-api-access-lr5r8\") pod \"collect-profiles-29556495-sh8g8\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.484273 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:00 crc kubenswrapper[4876]: I0313 08:15:00.970931 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8"] Mar 13 08:15:00 crc kubenswrapper[4876]: W0313 08:15:00.975462 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda44c13b9_7f84_4443_ac9f_52687a97990f.slice/crio-d2dff33a32947c9620004c86b8780d914d8535ba72a894fd5d68ae0b99fc4728 WatchSource:0}: Error finding container d2dff33a32947c9620004c86b8780d914d8535ba72a894fd5d68ae0b99fc4728: Status 404 returned error can't find the container with id d2dff33a32947c9620004c86b8780d914d8535ba72a894fd5d68ae0b99fc4728 Mar 13 08:15:01 crc kubenswrapper[4876]: I0313 08:15:01.366861 4876 generic.go:334] "Generic (PLEG): container finished" podID="a44c13b9-7f84-4443-ac9f-52687a97990f" containerID="b331ad4d2bfae66ae66f528a4b3f28b2b780bc14fa9bfc5ece56bc13a2437533" exitCode=0 Mar 13 08:15:01 crc kubenswrapper[4876]: I0313 08:15:01.366944 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" event={"ID":"a44c13b9-7f84-4443-ac9f-52687a97990f","Type":"ContainerDied","Data":"b331ad4d2bfae66ae66f528a4b3f28b2b780bc14fa9bfc5ece56bc13a2437533"} Mar 13 08:15:01 crc kubenswrapper[4876]: I0313 08:15:01.367134 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" event={"ID":"a44c13b9-7f84-4443-ac9f-52687a97990f","Type":"ContainerStarted","Data":"d2dff33a32947c9620004c86b8780d914d8535ba72a894fd5d68ae0b99fc4728"} Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.694439 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.743997 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a44c13b9-7f84-4443-ac9f-52687a97990f-config-volume\") pod \"a44c13b9-7f84-4443-ac9f-52687a97990f\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.744100 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr5r8\" (UniqueName: \"kubernetes.io/projected/a44c13b9-7f84-4443-ac9f-52687a97990f-kube-api-access-lr5r8\") pod \"a44c13b9-7f84-4443-ac9f-52687a97990f\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.744155 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a44c13b9-7f84-4443-ac9f-52687a97990f-secret-volume\") pod \"a44c13b9-7f84-4443-ac9f-52687a97990f\" (UID: \"a44c13b9-7f84-4443-ac9f-52687a97990f\") " Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.744914 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a44c13b9-7f84-4443-ac9f-52687a97990f-config-volume" (OuterVolumeSpecName: "config-volume") pod "a44c13b9-7f84-4443-ac9f-52687a97990f" (UID: "a44c13b9-7f84-4443-ac9f-52687a97990f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.752074 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a44c13b9-7f84-4443-ac9f-52687a97990f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a44c13b9-7f84-4443-ac9f-52687a97990f" (UID: "a44c13b9-7f84-4443-ac9f-52687a97990f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.752142 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a44c13b9-7f84-4443-ac9f-52687a97990f-kube-api-access-lr5r8" (OuterVolumeSpecName: "kube-api-access-lr5r8") pod "a44c13b9-7f84-4443-ac9f-52687a97990f" (UID: "a44c13b9-7f84-4443-ac9f-52687a97990f"). InnerVolumeSpecName "kube-api-access-lr5r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.845663 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a44c13b9-7f84-4443-ac9f-52687a97990f-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.845702 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr5r8\" (UniqueName: \"kubernetes.io/projected/a44c13b9-7f84-4443-ac9f-52687a97990f-kube-api-access-lr5r8\") on node \"crc\" DevicePath \"\"" Mar 13 08:15:02 crc kubenswrapper[4876]: I0313 08:15:02.845713 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a44c13b9-7f84-4443-ac9f-52687a97990f-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 08:15:03 crc kubenswrapper[4876]: I0313 08:15:03.384322 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" event={"ID":"a44c13b9-7f84-4443-ac9f-52687a97990f","Type":"ContainerDied","Data":"d2dff33a32947c9620004c86b8780d914d8535ba72a894fd5d68ae0b99fc4728"} Mar 13 08:15:03 crc kubenswrapper[4876]: I0313 08:15:03.384369 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2dff33a32947c9620004c86b8780d914d8535ba72a894fd5d68ae0b99fc4728" Mar 13 08:15:03 crc kubenswrapper[4876]: I0313 08:15:03.384646 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556495-sh8g8" Mar 13 08:15:03 crc kubenswrapper[4876]: I0313 08:15:03.770509 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm"] Mar 13 08:15:03 crc kubenswrapper[4876]: I0313 08:15:03.778069 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556450-9j8xm"] Mar 13 08:15:05 crc kubenswrapper[4876]: I0313 08:15:05.046491 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efc8b898-1330-4381-b780-ee41a9514f05" path="/var/lib/kubelet/pods/efc8b898-1330-4381-b780-ee41a9514f05/volumes" Mar 13 08:15:22 crc kubenswrapper[4876]: I0313 08:15:22.843367 4876 scope.go:117] "RemoveContainer" containerID="181d010f2fcc26b78278f10c1875f24e976a91b573cefd134991bf5558ea2faa" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.142175 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556496-xsw4v"] Mar 13 08:16:00 crc kubenswrapper[4876]: E0313 08:16:00.143177 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a44c13b9-7f84-4443-ac9f-52687a97990f" containerName="collect-profiles" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.143196 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a44c13b9-7f84-4443-ac9f-52687a97990f" containerName="collect-profiles" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.143433 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a44c13b9-7f84-4443-ac9f-52687a97990f" containerName="collect-profiles" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.144198 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556496-xsw4v" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.146064 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.146890 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.147017 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.153212 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556496-xsw4v"] Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.300406 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkc7z\" (UniqueName: \"kubernetes.io/projected/35b8ae2c-4713-4cab-8118-0f5ef2df30d5-kube-api-access-pkc7z\") pod \"auto-csr-approver-29556496-xsw4v\" (UID: \"35b8ae2c-4713-4cab-8118-0f5ef2df30d5\") " pod="openshift-infra/auto-csr-approver-29556496-xsw4v" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.401799 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkc7z\" (UniqueName: \"kubernetes.io/projected/35b8ae2c-4713-4cab-8118-0f5ef2df30d5-kube-api-access-pkc7z\") pod \"auto-csr-approver-29556496-xsw4v\" (UID: \"35b8ae2c-4713-4cab-8118-0f5ef2df30d5\") " pod="openshift-infra/auto-csr-approver-29556496-xsw4v" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.423820 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkc7z\" (UniqueName: \"kubernetes.io/projected/35b8ae2c-4713-4cab-8118-0f5ef2df30d5-kube-api-access-pkc7z\") pod \"auto-csr-approver-29556496-xsw4v\" (UID: \"35b8ae2c-4713-4cab-8118-0f5ef2df30d5\") " pod="openshift-infra/auto-csr-approver-29556496-xsw4v" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.467629 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556496-xsw4v" Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.887120 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556496-xsw4v"] Mar 13 08:16:00 crc kubenswrapper[4876]: I0313 08:16:00.917083 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556496-xsw4v" event={"ID":"35b8ae2c-4713-4cab-8118-0f5ef2df30d5","Type":"ContainerStarted","Data":"bb2afea235b90935196397e371b05fd28bbe8dc271a03e06b7dd8c7e2077fb8f"} Mar 13 08:16:02 crc kubenswrapper[4876]: I0313 08:16:02.933088 4876 generic.go:334] "Generic (PLEG): container finished" podID="35b8ae2c-4713-4cab-8118-0f5ef2df30d5" containerID="5a437662c3ad3512f2c539e65fc43f38b07d2770206c2545cda4d07df85a0d1a" exitCode=0 Mar 13 08:16:02 crc kubenswrapper[4876]: I0313 08:16:02.933157 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556496-xsw4v" event={"ID":"35b8ae2c-4713-4cab-8118-0f5ef2df30d5","Type":"ContainerDied","Data":"5a437662c3ad3512f2c539e65fc43f38b07d2770206c2545cda4d07df85a0d1a"} Mar 13 08:16:04 crc kubenswrapper[4876]: I0313 08:16:04.335654 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556496-xsw4v" Mar 13 08:16:04 crc kubenswrapper[4876]: I0313 08:16:04.477701 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkc7z\" (UniqueName: \"kubernetes.io/projected/35b8ae2c-4713-4cab-8118-0f5ef2df30d5-kube-api-access-pkc7z\") pod \"35b8ae2c-4713-4cab-8118-0f5ef2df30d5\" (UID: \"35b8ae2c-4713-4cab-8118-0f5ef2df30d5\") " Mar 13 08:16:04 crc kubenswrapper[4876]: I0313 08:16:04.485460 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35b8ae2c-4713-4cab-8118-0f5ef2df30d5-kube-api-access-pkc7z" (OuterVolumeSpecName: "kube-api-access-pkc7z") pod "35b8ae2c-4713-4cab-8118-0f5ef2df30d5" (UID: "35b8ae2c-4713-4cab-8118-0f5ef2df30d5"). InnerVolumeSpecName "kube-api-access-pkc7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:16:04 crc kubenswrapper[4876]: I0313 08:16:04.579836 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkc7z\" (UniqueName: \"kubernetes.io/projected/35b8ae2c-4713-4cab-8118-0f5ef2df30d5-kube-api-access-pkc7z\") on node \"crc\" DevicePath \"\"" Mar 13 08:16:04 crc kubenswrapper[4876]: I0313 08:16:04.950167 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556496-xsw4v" event={"ID":"35b8ae2c-4713-4cab-8118-0f5ef2df30d5","Type":"ContainerDied","Data":"bb2afea235b90935196397e371b05fd28bbe8dc271a03e06b7dd8c7e2077fb8f"} Mar 13 08:16:04 crc kubenswrapper[4876]: I0313 08:16:04.950203 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb2afea235b90935196397e371b05fd28bbe8dc271a03e06b7dd8c7e2077fb8f" Mar 13 08:16:04 crc kubenswrapper[4876]: I0313 08:16:04.950224 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556496-xsw4v" Mar 13 08:16:05 crc kubenswrapper[4876]: I0313 08:16:05.417085 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556490-ghmfx"] Mar 13 08:16:05 crc kubenswrapper[4876]: I0313 08:16:05.427210 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556490-ghmfx"] Mar 13 08:16:07 crc kubenswrapper[4876]: I0313 08:16:07.046739 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc332f0-d43a-4853-8017-47f67a849f2f" path="/var/lib/kubelet/pods/afc332f0-d43a-4853-8017-47f67a849f2f/volumes" Mar 13 08:16:22 crc kubenswrapper[4876]: I0313 08:16:22.936263 4876 scope.go:117] "RemoveContainer" containerID="24cf542e7878ea6641223e58c113d0325a2281dfc91432f3ea4f68153daf7050" Mar 13 08:16:24 crc kubenswrapper[4876]: I0313 08:16:24.608034 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:16:24 crc kubenswrapper[4876]: I0313 08:16:24.608645 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.397999 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-klkvd"] Mar 13 08:16:50 crc kubenswrapper[4876]: E0313 08:16:50.399721 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35b8ae2c-4713-4cab-8118-0f5ef2df30d5" containerName="oc" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.399738 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="35b8ae2c-4713-4cab-8118-0f5ef2df30d5" containerName="oc" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.400180 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="35b8ae2c-4713-4cab-8118-0f5ef2df30d5" containerName="oc" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.404398 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.411562 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klkvd"] Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.468459 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-catalog-content\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.468549 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8std\" (UniqueName: \"kubernetes.io/projected/44dee036-b550-40ba-832c-287209569620-kube-api-access-k8std\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.468690 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-utilities\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.571114 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-catalog-content\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.571506 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8std\" (UniqueName: \"kubernetes.io/projected/44dee036-b550-40ba-832c-287209569620-kube-api-access-k8std\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.571603 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-catalog-content\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.571731 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-utilities\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.572123 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-utilities\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.590935 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l9khp"] Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.593352 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.599226 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8std\" (UniqueName: \"kubernetes.io/projected/44dee036-b550-40ba-832c-287209569620-kube-api-access-k8std\") pod \"community-operators-klkvd\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.604135 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l9khp"] Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.673961 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrfc8\" (UniqueName: \"kubernetes.io/projected/747ee085-72b1-43b3-bbaa-3df21799e4c3-kube-api-access-mrfc8\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.674042 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-utilities\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.674075 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-catalog-content\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.730880 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.776217 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrfc8\" (UniqueName: \"kubernetes.io/projected/747ee085-72b1-43b3-bbaa-3df21799e4c3-kube-api-access-mrfc8\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.776278 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-catalog-content\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.776300 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-utilities\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.776805 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-utilities\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.777073 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-catalog-content\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.798925 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrfc8\" (UniqueName: \"kubernetes.io/projected/747ee085-72b1-43b3-bbaa-3df21799e4c3-kube-api-access-mrfc8\") pod \"certified-operators-l9khp\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:50 crc kubenswrapper[4876]: I0313 08:16:50.964710 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:16:51 crc kubenswrapper[4876]: I0313 08:16:51.363720 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klkvd"] Mar 13 08:16:51 crc kubenswrapper[4876]: I0313 08:16:51.399349 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkvd" event={"ID":"44dee036-b550-40ba-832c-287209569620","Type":"ContainerStarted","Data":"879389ef9eaef90dc9aec9097bea705c771cd792617803275f77bcd584d6d6b1"} Mar 13 08:16:51 crc kubenswrapper[4876]: I0313 08:16:51.503630 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l9khp"] Mar 13 08:16:51 crc kubenswrapper[4876]: W0313 08:16:51.536896 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod747ee085_72b1_43b3_bbaa_3df21799e4c3.slice/crio-8f56c0f2d574e4a1b88c4284e97922e9171c7e15ad97f48cf806d442ff44a450 WatchSource:0}: Error finding container 8f56c0f2d574e4a1b88c4284e97922e9171c7e15ad97f48cf806d442ff44a450: Status 404 returned error can't find the container with id 8f56c0f2d574e4a1b88c4284e97922e9171c7e15ad97f48cf806d442ff44a450 Mar 13 08:16:52 crc kubenswrapper[4876]: I0313 08:16:52.411227 4876 generic.go:334] "Generic (PLEG): container finished" podID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerID="32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d" exitCode=0 Mar 13 08:16:52 crc kubenswrapper[4876]: I0313 08:16:52.411357 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9khp" event={"ID":"747ee085-72b1-43b3-bbaa-3df21799e4c3","Type":"ContainerDied","Data":"32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d"} Mar 13 08:16:52 crc kubenswrapper[4876]: I0313 08:16:52.411688 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9khp" event={"ID":"747ee085-72b1-43b3-bbaa-3df21799e4c3","Type":"ContainerStarted","Data":"8f56c0f2d574e4a1b88c4284e97922e9171c7e15ad97f48cf806d442ff44a450"} Mar 13 08:16:52 crc kubenswrapper[4876]: I0313 08:16:52.413326 4876 generic.go:334] "Generic (PLEG): container finished" podID="44dee036-b550-40ba-832c-287209569620" containerID="a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6" exitCode=0 Mar 13 08:16:52 crc kubenswrapper[4876]: I0313 08:16:52.413377 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkvd" event={"ID":"44dee036-b550-40ba-832c-287209569620","Type":"ContainerDied","Data":"a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6"} Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.392857 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mgrwr"] Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.395700 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.410751 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgrwr"] Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.424403 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9khp" event={"ID":"747ee085-72b1-43b3-bbaa-3df21799e4c3","Type":"ContainerStarted","Data":"3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6"} Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.428060 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkvd" event={"ID":"44dee036-b550-40ba-832c-287209569620","Type":"ContainerStarted","Data":"130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe"} Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.436282 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-catalog-content\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.436517 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-utilities\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.436610 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf7nx\" (UniqueName: \"kubernetes.io/projected/2703705e-195d-4fcb-a966-8b04761aa0c1-kube-api-access-xf7nx\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.537988 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf7nx\" (UniqueName: \"kubernetes.io/projected/2703705e-195d-4fcb-a966-8b04761aa0c1-kube-api-access-xf7nx\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.538059 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-catalog-content\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.538228 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-utilities\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.538814 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-catalog-content\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.538843 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-utilities\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.566226 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf7nx\" (UniqueName: \"kubernetes.io/projected/2703705e-195d-4fcb-a966-8b04761aa0c1-kube-api-access-xf7nx\") pod \"redhat-marketplace-mgrwr\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:53 crc kubenswrapper[4876]: I0313 08:16:53.718355 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:16:54 crc kubenswrapper[4876]: W0313 08:16:54.183510 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2703705e_195d_4fcb_a966_8b04761aa0c1.slice/crio-78cebe130623a3b807df24a0b1a23e2cabca95b109e84ce6366fc8b7346c8f98 WatchSource:0}: Error finding container 78cebe130623a3b807df24a0b1a23e2cabca95b109e84ce6366fc8b7346c8f98: Status 404 returned error can't find the container with id 78cebe130623a3b807df24a0b1a23e2cabca95b109e84ce6366fc8b7346c8f98 Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.192744 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgrwr"] Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.438036 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgrwr" event={"ID":"2703705e-195d-4fcb-a966-8b04761aa0c1","Type":"ContainerStarted","Data":"944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3"} Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.438379 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgrwr" event={"ID":"2703705e-195d-4fcb-a966-8b04761aa0c1","Type":"ContainerStarted","Data":"78cebe130623a3b807df24a0b1a23e2cabca95b109e84ce6366fc8b7346c8f98"} Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.441827 4876 generic.go:334] "Generic (PLEG): container finished" podID="44dee036-b550-40ba-832c-287209569620" containerID="130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe" exitCode=0 Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.441887 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkvd" event={"ID":"44dee036-b550-40ba-832c-287209569620","Type":"ContainerDied","Data":"130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe"} Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.443672 4876 generic.go:334] "Generic (PLEG): container finished" podID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerID="3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6" exitCode=0 Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.443715 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9khp" event={"ID":"747ee085-72b1-43b3-bbaa-3df21799e4c3","Type":"ContainerDied","Data":"3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6"} Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.608207 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:16:54 crc kubenswrapper[4876]: I0313 08:16:54.608290 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:16:55 crc kubenswrapper[4876]: I0313 08:16:55.453620 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkvd" event={"ID":"44dee036-b550-40ba-832c-287209569620","Type":"ContainerStarted","Data":"76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d"} Mar 13 08:16:55 crc kubenswrapper[4876]: I0313 08:16:55.456312 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9khp" event={"ID":"747ee085-72b1-43b3-bbaa-3df21799e4c3","Type":"ContainerStarted","Data":"570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496"} Mar 13 08:16:55 crc kubenswrapper[4876]: I0313 08:16:55.458225 4876 generic.go:334] "Generic (PLEG): container finished" podID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerID="944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3" exitCode=0 Mar 13 08:16:55 crc kubenswrapper[4876]: I0313 08:16:55.458287 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgrwr" event={"ID":"2703705e-195d-4fcb-a966-8b04761aa0c1","Type":"ContainerDied","Data":"944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3"} Mar 13 08:16:55 crc kubenswrapper[4876]: I0313 08:16:55.494200 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-klkvd" podStartSLOduration=3.035282007 podStartE2EDuration="5.49417512s" podCreationTimestamp="2026-03-13 08:16:50 +0000 UTC" firstStartedPulling="2026-03-13 08:16:52.414860565 +0000 UTC m=+2272.085639537" lastFinishedPulling="2026-03-13 08:16:54.873753668 +0000 UTC m=+2274.544532650" observedRunningTime="2026-03-13 08:16:55.474731971 +0000 UTC m=+2275.145510953" watchObservedRunningTime="2026-03-13 08:16:55.49417512 +0000 UTC m=+2275.164954102" Mar 13 08:16:55 crc kubenswrapper[4876]: I0313 08:16:55.498975 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l9khp" podStartSLOduration=3.034885947 podStartE2EDuration="5.498953666s" podCreationTimestamp="2026-03-13 08:16:50 +0000 UTC" firstStartedPulling="2026-03-13 08:16:52.414861765 +0000 UTC m=+2272.085640747" lastFinishedPulling="2026-03-13 08:16:54.878929474 +0000 UTC m=+2274.549708466" observedRunningTime="2026-03-13 08:16:55.492606539 +0000 UTC m=+2275.163385521" watchObservedRunningTime="2026-03-13 08:16:55.498953666 +0000 UTC m=+2275.169732648" Mar 13 08:16:56 crc kubenswrapper[4876]: I0313 08:16:56.470285 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgrwr" event={"ID":"2703705e-195d-4fcb-a966-8b04761aa0c1","Type":"ContainerStarted","Data":"46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0"} Mar 13 08:16:57 crc kubenswrapper[4876]: I0313 08:16:57.481944 4876 generic.go:334] "Generic (PLEG): container finished" podID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerID="46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0" exitCode=0 Mar 13 08:16:57 crc kubenswrapper[4876]: I0313 08:16:57.482118 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgrwr" event={"ID":"2703705e-195d-4fcb-a966-8b04761aa0c1","Type":"ContainerDied","Data":"46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0"} Mar 13 08:16:58 crc kubenswrapper[4876]: I0313 08:16:58.500692 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgrwr" event={"ID":"2703705e-195d-4fcb-a966-8b04761aa0c1","Type":"ContainerStarted","Data":"328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa"} Mar 13 08:16:58 crc kubenswrapper[4876]: I0313 08:16:58.521192 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mgrwr" podStartSLOduration=2.852184913 podStartE2EDuration="5.521172277s" podCreationTimestamp="2026-03-13 08:16:53 +0000 UTC" firstStartedPulling="2026-03-13 08:16:55.459505683 +0000 UTC m=+2275.130284675" lastFinishedPulling="2026-03-13 08:16:58.128493057 +0000 UTC m=+2277.799272039" observedRunningTime="2026-03-13 08:16:58.519549744 +0000 UTC m=+2278.190328756" watchObservedRunningTime="2026-03-13 08:16:58.521172277 +0000 UTC m=+2278.191951259" Mar 13 08:17:00 crc kubenswrapper[4876]: I0313 08:17:00.731554 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:17:00 crc kubenswrapper[4876]: I0313 08:17:00.732012 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:17:00 crc kubenswrapper[4876]: I0313 08:17:00.966867 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:17:00 crc kubenswrapper[4876]: I0313 08:17:00.967257 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:17:01 crc kubenswrapper[4876]: I0313 08:17:01.008624 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:17:01 crc kubenswrapper[4876]: I0313 08:17:01.568950 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:17:01 crc kubenswrapper[4876]: I0313 08:17:01.780256 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-klkvd" podUID="44dee036-b550-40ba-832c-287209569620" containerName="registry-server" probeResult="failure" output=< Mar 13 08:17:01 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:17:01 crc kubenswrapper[4876]: > Mar 13 08:17:01 crc kubenswrapper[4876]: I0313 08:17:01.977625 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l9khp"] Mar 13 08:17:03 crc kubenswrapper[4876]: I0313 08:17:03.542551 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l9khp" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerName="registry-server" containerID="cri-o://570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496" gracePeriod=2 Mar 13 08:17:03 crc kubenswrapper[4876]: I0313 08:17:03.718806 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:17:03 crc kubenswrapper[4876]: I0313 08:17:03.718852 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:17:03 crc kubenswrapper[4876]: I0313 08:17:03.779555 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.006452 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.057317 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-utilities\") pod \"747ee085-72b1-43b3-bbaa-3df21799e4c3\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.057392 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-catalog-content\") pod \"747ee085-72b1-43b3-bbaa-3df21799e4c3\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.057501 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrfc8\" (UniqueName: \"kubernetes.io/projected/747ee085-72b1-43b3-bbaa-3df21799e4c3-kube-api-access-mrfc8\") pod \"747ee085-72b1-43b3-bbaa-3df21799e4c3\" (UID: \"747ee085-72b1-43b3-bbaa-3df21799e4c3\") " Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.058274 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-utilities" (OuterVolumeSpecName: "utilities") pod "747ee085-72b1-43b3-bbaa-3df21799e4c3" (UID: "747ee085-72b1-43b3-bbaa-3df21799e4c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.063282 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/747ee085-72b1-43b3-bbaa-3df21799e4c3-kube-api-access-mrfc8" (OuterVolumeSpecName: "kube-api-access-mrfc8") pod "747ee085-72b1-43b3-bbaa-3df21799e4c3" (UID: "747ee085-72b1-43b3-bbaa-3df21799e4c3"). InnerVolumeSpecName "kube-api-access-mrfc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.108687 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "747ee085-72b1-43b3-bbaa-3df21799e4c3" (UID: "747ee085-72b1-43b3-bbaa-3df21799e4c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.159112 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.159153 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrfc8\" (UniqueName: \"kubernetes.io/projected/747ee085-72b1-43b3-bbaa-3df21799e4c3-kube-api-access-mrfc8\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.159172 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747ee085-72b1-43b3-bbaa-3df21799e4c3-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.567078 4876 generic.go:334] "Generic (PLEG): container finished" podID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerID="570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496" exitCode=0 Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.567176 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9khp" event={"ID":"747ee085-72b1-43b3-bbaa-3df21799e4c3","Type":"ContainerDied","Data":"570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496"} Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.568104 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9khp" event={"ID":"747ee085-72b1-43b3-bbaa-3df21799e4c3","Type":"ContainerDied","Data":"8f56c0f2d574e4a1b88c4284e97922e9171c7e15ad97f48cf806d442ff44a450"} Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.568138 4876 scope.go:117] "RemoveContainer" containerID="570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.567275 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9khp" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.591542 4876 scope.go:117] "RemoveContainer" containerID="3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.607898 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l9khp"] Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.615350 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l9khp"] Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.617459 4876 scope.go:117] "RemoveContainer" containerID="32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.630457 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.669225 4876 scope.go:117] "RemoveContainer" containerID="570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496" Mar 13 08:17:04 crc kubenswrapper[4876]: E0313 08:17:04.669810 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496\": container with ID starting with 570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496 not found: ID does not exist" containerID="570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.669855 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496"} err="failed to get container status \"570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496\": rpc error: code = NotFound desc = could not find container \"570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496\": container with ID starting with 570317edfd2c23a905f27791ec732eb5d523d45d3d04dff42f103ba6974d9496 not found: ID does not exist" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.669883 4876 scope.go:117] "RemoveContainer" containerID="3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6" Mar 13 08:17:04 crc kubenswrapper[4876]: E0313 08:17:04.670189 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6\": container with ID starting with 3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6 not found: ID does not exist" containerID="3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.670232 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6"} err="failed to get container status \"3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6\": rpc error: code = NotFound desc = could not find container \"3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6\": container with ID starting with 3445c9520df25e02446e7e6dc77c77d5bc33e34b869b773ab71e04b09b5c6bd6 not found: ID does not exist" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.670272 4876 scope.go:117] "RemoveContainer" containerID="32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d" Mar 13 08:17:04 crc kubenswrapper[4876]: E0313 08:17:04.670693 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d\": container with ID starting with 32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d not found: ID does not exist" containerID="32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d" Mar 13 08:17:04 crc kubenswrapper[4876]: I0313 08:17:04.670721 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d"} err="failed to get container status \"32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d\": rpc error: code = NotFound desc = could not find container \"32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d\": container with ID starting with 32340c601243ee964483f7fba176823a5eb6360c606527971ad2a4cf2554b35d not found: ID does not exist" Mar 13 08:17:05 crc kubenswrapper[4876]: I0313 08:17:05.049641 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" path="/var/lib/kubelet/pods/747ee085-72b1-43b3-bbaa-3df21799e4c3/volumes" Mar 13 08:17:06 crc kubenswrapper[4876]: I0313 08:17:06.178001 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgrwr"] Mar 13 08:17:06 crc kubenswrapper[4876]: I0313 08:17:06.583790 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mgrwr" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerName="registry-server" containerID="cri-o://328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa" gracePeriod=2 Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.143573 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.224623 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-catalog-content\") pod \"2703705e-195d-4fcb-a966-8b04761aa0c1\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.224676 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf7nx\" (UniqueName: \"kubernetes.io/projected/2703705e-195d-4fcb-a966-8b04761aa0c1-kube-api-access-xf7nx\") pod \"2703705e-195d-4fcb-a966-8b04761aa0c1\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.224859 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-utilities\") pod \"2703705e-195d-4fcb-a966-8b04761aa0c1\" (UID: \"2703705e-195d-4fcb-a966-8b04761aa0c1\") " Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.226056 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-utilities" (OuterVolumeSpecName: "utilities") pod "2703705e-195d-4fcb-a966-8b04761aa0c1" (UID: "2703705e-195d-4fcb-a966-8b04761aa0c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.237595 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2703705e-195d-4fcb-a966-8b04761aa0c1-kube-api-access-xf7nx" (OuterVolumeSpecName: "kube-api-access-xf7nx") pod "2703705e-195d-4fcb-a966-8b04761aa0c1" (UID: "2703705e-195d-4fcb-a966-8b04761aa0c1"). InnerVolumeSpecName "kube-api-access-xf7nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.249858 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2703705e-195d-4fcb-a966-8b04761aa0c1" (UID: "2703705e-195d-4fcb-a966-8b04761aa0c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.327529 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.327574 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2703705e-195d-4fcb-a966-8b04761aa0c1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.327589 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf7nx\" (UniqueName: \"kubernetes.io/projected/2703705e-195d-4fcb-a966-8b04761aa0c1-kube-api-access-xf7nx\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.596261 4876 generic.go:334] "Generic (PLEG): container finished" podID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerID="328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa" exitCode=0 Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.596320 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgrwr" event={"ID":"2703705e-195d-4fcb-a966-8b04761aa0c1","Type":"ContainerDied","Data":"328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa"} Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.596356 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgrwr" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.596677 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgrwr" event={"ID":"2703705e-195d-4fcb-a966-8b04761aa0c1","Type":"ContainerDied","Data":"78cebe130623a3b807df24a0b1a23e2cabca95b109e84ce6366fc8b7346c8f98"} Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.596706 4876 scope.go:117] "RemoveContainer" containerID="328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.636677 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgrwr"] Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.636680 4876 scope.go:117] "RemoveContainer" containerID="46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.644563 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgrwr"] Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.661101 4876 scope.go:117] "RemoveContainer" containerID="944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.698076 4876 scope.go:117] "RemoveContainer" containerID="328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa" Mar 13 08:17:07 crc kubenswrapper[4876]: E0313 08:17:07.698530 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa\": container with ID starting with 328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa not found: ID does not exist" containerID="328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.698572 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa"} err="failed to get container status \"328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa\": rpc error: code = NotFound desc = could not find container \"328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa\": container with ID starting with 328a7165ced32523fe530fec415f7379517c8b2e913238e74d41a19fbb4cb8aa not found: ID does not exist" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.698615 4876 scope.go:117] "RemoveContainer" containerID="46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0" Mar 13 08:17:07 crc kubenswrapper[4876]: E0313 08:17:07.700588 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0\": container with ID starting with 46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0 not found: ID does not exist" containerID="46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.700621 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0"} err="failed to get container status \"46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0\": rpc error: code = NotFound desc = could not find container \"46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0\": container with ID starting with 46f44aa1a898749dff2dd752232a5bbad6c8ca129a82f271a6508c62e2ab3fe0 not found: ID does not exist" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.700645 4876 scope.go:117] "RemoveContainer" containerID="944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3" Mar 13 08:17:07 crc kubenswrapper[4876]: E0313 08:17:07.700867 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3\": container with ID starting with 944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3 not found: ID does not exist" containerID="944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3" Mar 13 08:17:07 crc kubenswrapper[4876]: I0313 08:17:07.700926 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3"} err="failed to get container status \"944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3\": rpc error: code = NotFound desc = could not find container \"944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3\": container with ID starting with 944a71ad99b8db4d2b1cb94436f015e91ac3df9975579424ebc52d91fc4d8aa3 not found: ID does not exist" Mar 13 08:17:09 crc kubenswrapper[4876]: I0313 08:17:09.049679 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" path="/var/lib/kubelet/pods/2703705e-195d-4fcb-a966-8b04761aa0c1/volumes" Mar 13 08:17:10 crc kubenswrapper[4876]: I0313 08:17:10.794161 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:17:10 crc kubenswrapper[4876]: I0313 08:17:10.870982 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:17:11 crc kubenswrapper[4876]: I0313 08:17:11.588134 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klkvd"] Mar 13 08:17:12 crc kubenswrapper[4876]: I0313 08:17:12.647832 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-klkvd" podUID="44dee036-b550-40ba-832c-287209569620" containerName="registry-server" containerID="cri-o://76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d" gracePeriod=2 Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.103385 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.159389 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-catalog-content\") pod \"44dee036-b550-40ba-832c-287209569620\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.159786 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-utilities\") pod \"44dee036-b550-40ba-832c-287209569620\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.159981 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8std\" (UniqueName: \"kubernetes.io/projected/44dee036-b550-40ba-832c-287209569620-kube-api-access-k8std\") pod \"44dee036-b550-40ba-832c-287209569620\" (UID: \"44dee036-b550-40ba-832c-287209569620\") " Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.161751 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-utilities" (OuterVolumeSpecName: "utilities") pod "44dee036-b550-40ba-832c-287209569620" (UID: "44dee036-b550-40ba-832c-287209569620"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.170304 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44dee036-b550-40ba-832c-287209569620-kube-api-access-k8std" (OuterVolumeSpecName: "kube-api-access-k8std") pod "44dee036-b550-40ba-832c-287209569620" (UID: "44dee036-b550-40ba-832c-287209569620"). InnerVolumeSpecName "kube-api-access-k8std". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.220468 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44dee036-b550-40ba-832c-287209569620" (UID: "44dee036-b550-40ba-832c-287209569620"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.262232 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8std\" (UniqueName: \"kubernetes.io/projected/44dee036-b550-40ba-832c-287209569620-kube-api-access-k8std\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.262286 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.262296 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44dee036-b550-40ba-832c-287209569620-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.661853 4876 generic.go:334] "Generic (PLEG): container finished" podID="44dee036-b550-40ba-832c-287209569620" containerID="76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d" exitCode=0 Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.661905 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkvd" event={"ID":"44dee036-b550-40ba-832c-287209569620","Type":"ContainerDied","Data":"76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d"} Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.661984 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klkvd" event={"ID":"44dee036-b550-40ba-832c-287209569620","Type":"ContainerDied","Data":"879389ef9eaef90dc9aec9097bea705c771cd792617803275f77bcd584d6d6b1"} Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.662013 4876 scope.go:117] "RemoveContainer" containerID="76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.661933 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klkvd" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.693691 4876 scope.go:117] "RemoveContainer" containerID="130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.701311 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klkvd"] Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.715069 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-klkvd"] Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.726435 4876 scope.go:117] "RemoveContainer" containerID="a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.779679 4876 scope.go:117] "RemoveContainer" containerID="76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d" Mar 13 08:17:13 crc kubenswrapper[4876]: E0313 08:17:13.780140 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d\": container with ID starting with 76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d not found: ID does not exist" containerID="76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.780170 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d"} err="failed to get container status \"76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d\": rpc error: code = NotFound desc = could not find container \"76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d\": container with ID starting with 76e63b20c357dd4184aad37cb54701314b82e9212d73061ddc5670222b0c792d not found: ID does not exist" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.780189 4876 scope.go:117] "RemoveContainer" containerID="130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe" Mar 13 08:17:13 crc kubenswrapper[4876]: E0313 08:17:13.780510 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe\": container with ID starting with 130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe not found: ID does not exist" containerID="130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.780531 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe"} err="failed to get container status \"130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe\": rpc error: code = NotFound desc = could not find container \"130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe\": container with ID starting with 130a5f3cf38f18757885ffea7693800acaee1428443928ae32d6fd31e0b0e2fe not found: ID does not exist" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.780542 4876 scope.go:117] "RemoveContainer" containerID="a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6" Mar 13 08:17:13 crc kubenswrapper[4876]: E0313 08:17:13.780752 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6\": container with ID starting with a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6 not found: ID does not exist" containerID="a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6" Mar 13 08:17:13 crc kubenswrapper[4876]: I0313 08:17:13.780772 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6"} err="failed to get container status \"a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6\": rpc error: code = NotFound desc = could not find container \"a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6\": container with ID starting with a6deb4996fdb806373cc2c502a230a9d48f992ff2d45d0a5ea19e74188d6aac6 not found: ID does not exist" Mar 13 08:17:15 crc kubenswrapper[4876]: I0313 08:17:15.048126 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44dee036-b550-40ba-832c-287209569620" path="/var/lib/kubelet/pods/44dee036-b550-40ba-832c-287209569620/volumes" Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.607299 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.607761 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.607822 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.608594 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.608649 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" gracePeriod=600 Mar 13 08:17:24 crc kubenswrapper[4876]: E0313 08:17:24.733146 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.764996 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" exitCode=0 Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.765041 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66"} Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.765076 4876 scope.go:117] "RemoveContainer" containerID="d0c73df6ed9f0626ecd967ee4e119fdb9bc95174ba5b98bf45405cb2d0c94943" Mar 13 08:17:24 crc kubenswrapper[4876]: I0313 08:17:24.766734 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:17:24 crc kubenswrapper[4876]: E0313 08:17:24.767086 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:17:36 crc kubenswrapper[4876]: I0313 08:17:36.036301 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:17:36 crc kubenswrapper[4876]: E0313 08:17:36.037982 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:17:49 crc kubenswrapper[4876]: I0313 08:17:49.036311 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:17:49 crc kubenswrapper[4876]: E0313 08:17:49.037016 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.163021 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556498-mv7bp"] Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.163926 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerName="extract-content" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.163938 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerName="extract-content" Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.163951 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerName="extract-utilities" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.163956 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerName="extract-utilities" Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.163969 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.163974 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.163984 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dee036-b550-40ba-832c-287209569620" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.163990 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dee036-b550-40ba-832c-287209569620" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.164001 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerName="extract-content" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164008 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerName="extract-content" Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.164021 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerName="extract-utilities" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164026 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerName="extract-utilities" Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.164035 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dee036-b550-40ba-832c-287209569620" containerName="extract-utilities" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164040 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dee036-b550-40ba-832c-287209569620" containerName="extract-utilities" Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.164051 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44dee036-b550-40ba-832c-287209569620" containerName="extract-content" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164056 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44dee036-b550-40ba-832c-287209569620" containerName="extract-content" Mar 13 08:18:00 crc kubenswrapper[4876]: E0313 08:18:00.164065 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164070 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164227 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="747ee085-72b1-43b3-bbaa-3df21799e4c3" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164315 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2703705e-195d-4fcb-a966-8b04761aa0c1" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164328 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="44dee036-b550-40ba-832c-287209569620" containerName="registry-server" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.164902 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556498-mv7bp" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.167086 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.167338 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.167359 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.183158 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556498-mv7bp"] Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.213384 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbrgm\" (UniqueName: \"kubernetes.io/projected/80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5-kube-api-access-fbrgm\") pod \"auto-csr-approver-29556498-mv7bp\" (UID: \"80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5\") " pod="openshift-infra/auto-csr-approver-29556498-mv7bp" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.315759 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbrgm\" (UniqueName: \"kubernetes.io/projected/80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5-kube-api-access-fbrgm\") pod \"auto-csr-approver-29556498-mv7bp\" (UID: \"80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5\") " pod="openshift-infra/auto-csr-approver-29556498-mv7bp" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.335771 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbrgm\" (UniqueName: \"kubernetes.io/projected/80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5-kube-api-access-fbrgm\") pod \"auto-csr-approver-29556498-mv7bp\" (UID: \"80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5\") " pod="openshift-infra/auto-csr-approver-29556498-mv7bp" Mar 13 08:18:00 crc kubenswrapper[4876]: I0313 08:18:00.487804 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556498-mv7bp" Mar 13 08:18:01 crc kubenswrapper[4876]: I0313 08:18:01.082030 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556498-mv7bp"] Mar 13 08:18:01 crc kubenswrapper[4876]: I0313 08:18:01.109882 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556498-mv7bp" event={"ID":"80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5","Type":"ContainerStarted","Data":"466ec321aee57deb3d9732f355d1ece800d8dc0888195995d73c0e2983f3096f"} Mar 13 08:18:03 crc kubenswrapper[4876]: I0313 08:18:03.035666 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:18:03 crc kubenswrapper[4876]: E0313 08:18:03.036758 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:18:03 crc kubenswrapper[4876]: I0313 08:18:03.127353 4876 generic.go:334] "Generic (PLEG): container finished" podID="80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5" containerID="6bfcf56cb337dd671211118065de14c3cddbe0810628f832b62b06a1648733ed" exitCode=0 Mar 13 08:18:03 crc kubenswrapper[4876]: I0313 08:18:03.127397 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556498-mv7bp" event={"ID":"80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5","Type":"ContainerDied","Data":"6bfcf56cb337dd671211118065de14c3cddbe0810628f832b62b06a1648733ed"} Mar 13 08:18:04 crc kubenswrapper[4876]: I0313 08:18:04.536831 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556498-mv7bp" Mar 13 08:18:04 crc kubenswrapper[4876]: I0313 08:18:04.612959 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbrgm\" (UniqueName: \"kubernetes.io/projected/80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5-kube-api-access-fbrgm\") pod \"80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5\" (UID: \"80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5\") " Mar 13 08:18:04 crc kubenswrapper[4876]: I0313 08:18:04.618720 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5-kube-api-access-fbrgm" (OuterVolumeSpecName: "kube-api-access-fbrgm") pod "80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5" (UID: "80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5"). InnerVolumeSpecName "kube-api-access-fbrgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:18:04 crc kubenswrapper[4876]: I0313 08:18:04.715158 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbrgm\" (UniqueName: \"kubernetes.io/projected/80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5-kube-api-access-fbrgm\") on node \"crc\" DevicePath \"\"" Mar 13 08:18:05 crc kubenswrapper[4876]: I0313 08:18:05.148880 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556498-mv7bp" event={"ID":"80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5","Type":"ContainerDied","Data":"466ec321aee57deb3d9732f355d1ece800d8dc0888195995d73c0e2983f3096f"} Mar 13 08:18:05 crc kubenswrapper[4876]: I0313 08:18:05.148962 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="466ec321aee57deb3d9732f355d1ece800d8dc0888195995d73c0e2983f3096f" Mar 13 08:18:05 crc kubenswrapper[4876]: I0313 08:18:05.149015 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556498-mv7bp" Mar 13 08:18:05 crc kubenswrapper[4876]: I0313 08:18:05.604807 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556492-x58br"] Mar 13 08:18:05 crc kubenswrapper[4876]: I0313 08:18:05.612436 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556492-x58br"] Mar 13 08:18:07 crc kubenswrapper[4876]: I0313 08:18:07.060963 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="272eadd8-ddf0-44c4-8daa-0dcbc371aa6a" path="/var/lib/kubelet/pods/272eadd8-ddf0-44c4-8daa-0dcbc371aa6a/volumes" Mar 13 08:18:14 crc kubenswrapper[4876]: I0313 08:18:14.036102 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:18:14 crc kubenswrapper[4876]: E0313 08:18:14.036872 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:18:18 crc kubenswrapper[4876]: I0313 08:18:18.259684 4876 generic.go:334] "Generic (PLEG): container finished" podID="240f0880-695b-44b4-ae52-2beacd34f2d0" containerID="6a01fecc812f6430ee52ed38fa49922eae8e8e2061980b991f2cb03a4ba9c28e" exitCode=0 Mar 13 08:18:18 crc kubenswrapper[4876]: I0313 08:18:18.259784 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" event={"ID":"240f0880-695b-44b4-ae52-2beacd34f2d0","Type":"ContainerDied","Data":"6a01fecc812f6430ee52ed38fa49922eae8e8e2061980b991f2cb03a4ba9c28e"} Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.708347 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.720455 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-secret-0\") pod \"240f0880-695b-44b4-ae52-2beacd34f2d0\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.720538 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-inventory\") pod \"240f0880-695b-44b4-ae52-2beacd34f2d0\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.720704 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-combined-ca-bundle\") pod \"240f0880-695b-44b4-ae52-2beacd34f2d0\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.720747 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7pnj\" (UniqueName: \"kubernetes.io/projected/240f0880-695b-44b4-ae52-2beacd34f2d0-kube-api-access-p7pnj\") pod \"240f0880-695b-44b4-ae52-2beacd34f2d0\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.720862 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-ssh-key-openstack-edpm-ipam\") pod \"240f0880-695b-44b4-ae52-2beacd34f2d0\" (UID: \"240f0880-695b-44b4-ae52-2beacd34f2d0\") " Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.728418 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "240f0880-695b-44b4-ae52-2beacd34f2d0" (UID: "240f0880-695b-44b4-ae52-2beacd34f2d0"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.730960 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/240f0880-695b-44b4-ae52-2beacd34f2d0-kube-api-access-p7pnj" (OuterVolumeSpecName: "kube-api-access-p7pnj") pod "240f0880-695b-44b4-ae52-2beacd34f2d0" (UID: "240f0880-695b-44b4-ae52-2beacd34f2d0"). InnerVolumeSpecName "kube-api-access-p7pnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.756262 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-inventory" (OuterVolumeSpecName: "inventory") pod "240f0880-695b-44b4-ae52-2beacd34f2d0" (UID: "240f0880-695b-44b4-ae52-2beacd34f2d0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.766335 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "240f0880-695b-44b4-ae52-2beacd34f2d0" (UID: "240f0880-695b-44b4-ae52-2beacd34f2d0"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.769449 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "240f0880-695b-44b4-ae52-2beacd34f2d0" (UID: "240f0880-695b-44b4-ae52-2beacd34f2d0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.823427 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.823707 4876 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.823800 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.823885 4876 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/240f0880-695b-44b4-ae52-2beacd34f2d0-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:18:19 crc kubenswrapper[4876]: I0313 08:18:19.823958 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7pnj\" (UniqueName: \"kubernetes.io/projected/240f0880-695b-44b4-ae52-2beacd34f2d0-kube-api-access-p7pnj\") on node \"crc\" DevicePath \"\"" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.280661 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" event={"ID":"240f0880-695b-44b4-ae52-2beacd34f2d0","Type":"ContainerDied","Data":"3cec191ec1de73416f5f6ec968fdb3a267722005b7b91f3eea9a8a3ac761480f"} Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.280728 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cec191ec1de73416f5f6ec968fdb3a267722005b7b91f3eea9a8a3ac761480f" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.280698 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.364597 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4"] Mar 13 08:18:20 crc kubenswrapper[4876]: E0313 08:18:20.365026 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5" containerName="oc" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.365045 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5" containerName="oc" Mar 13 08:18:20 crc kubenswrapper[4876]: E0313 08:18:20.365082 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="240f0880-695b-44b4-ae52-2beacd34f2d0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.365094 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="240f0880-695b-44b4-ae52-2beacd34f2d0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.365331 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="240f0880-695b-44b4-ae52-2beacd34f2d0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.365362 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5" containerName="oc" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.366120 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.368400 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.369213 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.372892 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.373093 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.373278 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.373386 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.375090 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.379836 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4"] Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.435862 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.436286 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7b648455-e602-439e-b806-743233acb285-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.436432 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.436614 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr4b2\" (UniqueName: \"kubernetes.io/projected/7b648455-e602-439e-b806-743233acb285-kube-api-access-fr4b2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.436722 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.436845 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.436885 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.436940 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.437158 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.437334 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.437369 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539301 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr4b2\" (UniqueName: \"kubernetes.io/projected/7b648455-e602-439e-b806-743233acb285-kube-api-access-fr4b2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539361 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539432 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539462 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539494 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539530 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539599 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539624 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539657 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539695 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7b648455-e602-439e-b806-743233acb285-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.539738 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.540656 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7b648455-e602-439e-b806-743233acb285-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.542950 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.543203 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.543780 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.543872 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.544025 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.544522 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.545806 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.546747 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.547334 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.557143 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr4b2\" (UniqueName: \"kubernetes.io/projected/7b648455-e602-439e-b806-743233acb285-kube-api-access-fr4b2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-2xhz4\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:20 crc kubenswrapper[4876]: I0313 08:18:20.683347 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:18:21 crc kubenswrapper[4876]: I0313 08:18:21.228422 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4"] Mar 13 08:18:21 crc kubenswrapper[4876]: W0313 08:18:21.231830 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b648455_e602_439e_b806_743233acb285.slice/crio-0fccf9a88fc6439639ac8ae2c1b6ba35e481787aca88458de22a23f1ec77d911 WatchSource:0}: Error finding container 0fccf9a88fc6439639ac8ae2c1b6ba35e481787aca88458de22a23f1ec77d911: Status 404 returned error can't find the container with id 0fccf9a88fc6439639ac8ae2c1b6ba35e481787aca88458de22a23f1ec77d911 Mar 13 08:18:21 crc kubenswrapper[4876]: I0313 08:18:21.294189 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" event={"ID":"7b648455-e602-439e-b806-743233acb285","Type":"ContainerStarted","Data":"0fccf9a88fc6439639ac8ae2c1b6ba35e481787aca88458de22a23f1ec77d911"} Mar 13 08:18:22 crc kubenswrapper[4876]: I0313 08:18:22.307964 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" event={"ID":"7b648455-e602-439e-b806-743233acb285","Type":"ContainerStarted","Data":"7cec7b9b97bc635cff42fe271a82c11609851f08be05a227f6466afabd0e2dc1"} Mar 13 08:18:23 crc kubenswrapper[4876]: I0313 08:18:23.142851 4876 scope.go:117] "RemoveContainer" containerID="c1ed275ae00f56125e9ec6bdef2d336739470ed3222b67a1e1316622051726ec" Mar 13 08:18:27 crc kubenswrapper[4876]: I0313 08:18:27.036054 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:18:27 crc kubenswrapper[4876]: E0313 08:18:27.036993 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:18:41 crc kubenswrapper[4876]: I0313 08:18:41.049710 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:18:41 crc kubenswrapper[4876]: E0313 08:18:41.050652 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:18:53 crc kubenswrapper[4876]: I0313 08:18:53.035751 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:18:53 crc kubenswrapper[4876]: E0313 08:18:53.036650 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:19:05 crc kubenswrapper[4876]: I0313 08:19:05.036226 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:19:05 crc kubenswrapper[4876]: E0313 08:19:05.037114 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:19:18 crc kubenswrapper[4876]: I0313 08:19:18.035637 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:19:18 crc kubenswrapper[4876]: E0313 08:19:18.036454 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:19:32 crc kubenswrapper[4876]: I0313 08:19:32.036555 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:19:32 crc kubenswrapper[4876]: E0313 08:19:32.037475 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:19:43 crc kubenswrapper[4876]: I0313 08:19:43.053335 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:19:43 crc kubenswrapper[4876]: E0313 08:19:43.060198 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:19:57 crc kubenswrapper[4876]: I0313 08:19:57.035308 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:19:57 crc kubenswrapper[4876]: E0313 08:19:57.037807 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.159616 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" podStartSLOduration=99.557011541 podStartE2EDuration="1m40.159590065s" podCreationTimestamp="2026-03-13 08:18:20 +0000 UTC" firstStartedPulling="2026-03-13 08:18:21.235276432 +0000 UTC m=+2360.906055414" lastFinishedPulling="2026-03-13 08:18:21.837854936 +0000 UTC m=+2361.508633938" observedRunningTime="2026-03-13 08:18:22.328094341 +0000 UTC m=+2361.998873393" watchObservedRunningTime="2026-03-13 08:20:00.159590065 +0000 UTC m=+2459.830369057" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.166642 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556500-fg284"] Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.168390 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556500-fg284" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.170550 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.171020 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.171946 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.176163 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556500-fg284"] Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.281916 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm6d9\" (UniqueName: \"kubernetes.io/projected/5024f93c-6cfc-4019-9ad4-b39aee51d02e-kube-api-access-sm6d9\") pod \"auto-csr-approver-29556500-fg284\" (UID: \"5024f93c-6cfc-4019-9ad4-b39aee51d02e\") " pod="openshift-infra/auto-csr-approver-29556500-fg284" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.384700 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm6d9\" (UniqueName: \"kubernetes.io/projected/5024f93c-6cfc-4019-9ad4-b39aee51d02e-kube-api-access-sm6d9\") pod \"auto-csr-approver-29556500-fg284\" (UID: \"5024f93c-6cfc-4019-9ad4-b39aee51d02e\") " pod="openshift-infra/auto-csr-approver-29556500-fg284" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.402459 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm6d9\" (UniqueName: \"kubernetes.io/projected/5024f93c-6cfc-4019-9ad4-b39aee51d02e-kube-api-access-sm6d9\") pod \"auto-csr-approver-29556500-fg284\" (UID: \"5024f93c-6cfc-4019-9ad4-b39aee51d02e\") " pod="openshift-infra/auto-csr-approver-29556500-fg284" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.493948 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556500-fg284" Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.978770 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556500-fg284"] Mar 13 08:20:00 crc kubenswrapper[4876]: I0313 08:20:00.979663 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:20:01 crc kubenswrapper[4876]: I0313 08:20:01.276027 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556500-fg284" event={"ID":"5024f93c-6cfc-4019-9ad4-b39aee51d02e","Type":"ContainerStarted","Data":"5b2a20da047b0941caaaab23903802c324a25670cb05a71b5231f0f9bfb8d9f7"} Mar 13 08:20:02 crc kubenswrapper[4876]: I0313 08:20:02.287946 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556500-fg284" event={"ID":"5024f93c-6cfc-4019-9ad4-b39aee51d02e","Type":"ContainerStarted","Data":"7745192dd8ac3dab7eac784d445a7a890627ec3e88601669549e11c6fe1d0a07"} Mar 13 08:20:02 crc kubenswrapper[4876]: I0313 08:20:02.307445 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556500-fg284" podStartSLOduration=1.448244886 podStartE2EDuration="2.307428566s" podCreationTimestamp="2026-03-13 08:20:00 +0000 UTC" firstStartedPulling="2026-03-13 08:20:00.979315632 +0000 UTC m=+2460.650094624" lastFinishedPulling="2026-03-13 08:20:01.838499322 +0000 UTC m=+2461.509278304" observedRunningTime="2026-03-13 08:20:02.303020304 +0000 UTC m=+2461.973799326" watchObservedRunningTime="2026-03-13 08:20:02.307428566 +0000 UTC m=+2461.978207548" Mar 13 08:20:03 crc kubenswrapper[4876]: I0313 08:20:03.299669 4876 generic.go:334] "Generic (PLEG): container finished" podID="5024f93c-6cfc-4019-9ad4-b39aee51d02e" containerID="7745192dd8ac3dab7eac784d445a7a890627ec3e88601669549e11c6fe1d0a07" exitCode=0 Mar 13 08:20:03 crc kubenswrapper[4876]: I0313 08:20:03.299777 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556500-fg284" event={"ID":"5024f93c-6cfc-4019-9ad4-b39aee51d02e","Type":"ContainerDied","Data":"7745192dd8ac3dab7eac784d445a7a890627ec3e88601669549e11c6fe1d0a07"} Mar 13 08:20:04 crc kubenswrapper[4876]: I0313 08:20:04.654082 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556500-fg284" Mar 13 08:20:04 crc kubenswrapper[4876]: I0313 08:20:04.768132 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm6d9\" (UniqueName: \"kubernetes.io/projected/5024f93c-6cfc-4019-9ad4-b39aee51d02e-kube-api-access-sm6d9\") pod \"5024f93c-6cfc-4019-9ad4-b39aee51d02e\" (UID: \"5024f93c-6cfc-4019-9ad4-b39aee51d02e\") " Mar 13 08:20:04 crc kubenswrapper[4876]: I0313 08:20:04.775707 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5024f93c-6cfc-4019-9ad4-b39aee51d02e-kube-api-access-sm6d9" (OuterVolumeSpecName: "kube-api-access-sm6d9") pod "5024f93c-6cfc-4019-9ad4-b39aee51d02e" (UID: "5024f93c-6cfc-4019-9ad4-b39aee51d02e"). InnerVolumeSpecName "kube-api-access-sm6d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:20:04 crc kubenswrapper[4876]: I0313 08:20:04.870031 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm6d9\" (UniqueName: \"kubernetes.io/projected/5024f93c-6cfc-4019-9ad4-b39aee51d02e-kube-api-access-sm6d9\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:05 crc kubenswrapper[4876]: I0313 08:20:05.326623 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556500-fg284" event={"ID":"5024f93c-6cfc-4019-9ad4-b39aee51d02e","Type":"ContainerDied","Data":"5b2a20da047b0941caaaab23903802c324a25670cb05a71b5231f0f9bfb8d9f7"} Mar 13 08:20:05 crc kubenswrapper[4876]: I0313 08:20:05.326666 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b2a20da047b0941caaaab23903802c324a25670cb05a71b5231f0f9bfb8d9f7" Mar 13 08:20:05 crc kubenswrapper[4876]: I0313 08:20:05.326731 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556500-fg284" Mar 13 08:20:05 crc kubenswrapper[4876]: I0313 08:20:05.404545 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556494-69d6c"] Mar 13 08:20:05 crc kubenswrapper[4876]: I0313 08:20:05.412604 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556494-69d6c"] Mar 13 08:20:07 crc kubenswrapper[4876]: I0313 08:20:07.044960 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d16336-44b2-4142-a4e7-d0770f7a96cd" path="/var/lib/kubelet/pods/d9d16336-44b2-4142-a4e7-d0770f7a96cd/volumes" Mar 13 08:20:08 crc kubenswrapper[4876]: I0313 08:20:08.035565 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:20:08 crc kubenswrapper[4876]: E0313 08:20:08.036101 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:20:20 crc kubenswrapper[4876]: I0313 08:20:20.035728 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:20:20 crc kubenswrapper[4876]: E0313 08:20:20.036518 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:20:23 crc kubenswrapper[4876]: I0313 08:20:23.221912 4876 scope.go:117] "RemoveContainer" containerID="d42bb885c59a8700da179f0ae109e5eaa5f35cde19b9357d32fae60d0c9302f5" Mar 13 08:20:33 crc kubenswrapper[4876]: I0313 08:20:33.037695 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:20:33 crc kubenswrapper[4876]: E0313 08:20:33.038895 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:20:43 crc kubenswrapper[4876]: I0313 08:20:43.859579 4876 generic.go:334] "Generic (PLEG): container finished" podID="7b648455-e602-439e-b806-743233acb285" containerID="7cec7b9b97bc635cff42fe271a82c11609851f08be05a227f6466afabd0e2dc1" exitCode=0 Mar 13 08:20:43 crc kubenswrapper[4876]: I0313 08:20:43.859665 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" event={"ID":"7b648455-e602-439e-b806-743233acb285","Type":"ContainerDied","Data":"7cec7b9b97bc635cff42fe271a82c11609851f08be05a227f6466afabd0e2dc1"} Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.421903 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.449882 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-inventory\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.449951 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-3\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450020 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7b648455-e602-439e-b806-743233acb285-nova-extra-config-0\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450046 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-0\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450072 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-1\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450127 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-0\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450178 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-combined-ca-bundle\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450208 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-2\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450321 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-ssh-key-openstack-edpm-ipam\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450356 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr4b2\" (UniqueName: \"kubernetes.io/projected/7b648455-e602-439e-b806-743233acb285-kube-api-access-fr4b2\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.450396 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-1\") pod \"7b648455-e602-439e-b806-743233acb285\" (UID: \"7b648455-e602-439e-b806-743233acb285\") " Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.488464 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b648455-e602-439e-b806-743233acb285-kube-api-access-fr4b2" (OuterVolumeSpecName: "kube-api-access-fr4b2") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "kube-api-access-fr4b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.491508 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.507825 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.512383 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-inventory" (OuterVolumeSpecName: "inventory") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.547215 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.550298 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b648455-e602-439e-b806-743233acb285-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.552092 4876 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7b648455-e602-439e-b806-743233acb285-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.552121 4876 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.552133 4876 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.552142 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.552151 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr4b2\" (UniqueName: \"kubernetes.io/projected/7b648455-e602-439e-b806-743233acb285-kube-api-access-fr4b2\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.552161 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.572402 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.572442 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.579524 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.589574 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.593941 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "7b648455-e602-439e-b806-743233acb285" (UID: "7b648455-e602-439e-b806-743233acb285"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.653525 4876 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.653563 4876 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.653573 4876 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.653584 4876 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.653594 4876 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7b648455-e602-439e-b806-743233acb285-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.882137 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" event={"ID":"7b648455-e602-439e-b806-743233acb285","Type":"ContainerDied","Data":"0fccf9a88fc6439639ac8ae2c1b6ba35e481787aca88458de22a23f1ec77d911"} Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.882215 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fccf9a88fc6439639ac8ae2c1b6ba35e481787aca88458de22a23f1ec77d911" Mar 13 08:20:45 crc kubenswrapper[4876]: I0313 08:20:45.882260 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-2xhz4" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.031269 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx"] Mar 13 08:20:46 crc kubenswrapper[4876]: E0313 08:20:46.031742 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5024f93c-6cfc-4019-9ad4-b39aee51d02e" containerName="oc" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.031764 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5024f93c-6cfc-4019-9ad4-b39aee51d02e" containerName="oc" Mar 13 08:20:46 crc kubenswrapper[4876]: E0313 08:20:46.031804 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b648455-e602-439e-b806-743233acb285" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.031811 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b648455-e602-439e-b806-743233acb285" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.032037 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b648455-e602-439e-b806-743233acb285" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.032067 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5024f93c-6cfc-4019-9ad4-b39aee51d02e" containerName="oc" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.032781 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.036752 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:20:46 crc kubenswrapper[4876]: E0313 08:20:46.037187 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.038740 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.038932 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.039196 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.039428 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lwrcj" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.039215 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.062905 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.063067 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.063261 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.063323 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.063368 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.063581 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plcpn\" (UniqueName: \"kubernetes.io/projected/18be12a3-56e2-4b39-8c8f-784a7cafed2b-kube-api-access-plcpn\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.063634 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.064530 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx"] Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.165696 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.165858 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plcpn\" (UniqueName: \"kubernetes.io/projected/18be12a3-56e2-4b39-8c8f-784a7cafed2b-kube-api-access-plcpn\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.165896 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.165961 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.166003 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.166062 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.166108 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.170338 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.170396 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.171331 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.171416 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.171567 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.172321 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.188814 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plcpn\" (UniqueName: \"kubernetes.io/projected/18be12a3-56e2-4b39-8c8f-784a7cafed2b-kube-api-access-plcpn\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.368222 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.759561 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx"] Mar 13 08:20:46 crc kubenswrapper[4876]: I0313 08:20:46.896273 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" event={"ID":"18be12a3-56e2-4b39-8c8f-784a7cafed2b","Type":"ContainerStarted","Data":"08c80c63531bcd376f8616a1cfab25b4a0020e7ba8869d99643b5663a037a69a"} Mar 13 08:20:47 crc kubenswrapper[4876]: I0313 08:20:47.911230 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" event={"ID":"18be12a3-56e2-4b39-8c8f-784a7cafed2b","Type":"ContainerStarted","Data":"771e73610b3fd6e6da486126216413fc946ed3fd7e26b83b11a5a17cca9cf14e"} Mar 13 08:20:47 crc kubenswrapper[4876]: I0313 08:20:47.938383 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" podStartSLOduration=1.224793674 podStartE2EDuration="1.938352501s" podCreationTimestamp="2026-03-13 08:20:46 +0000 UTC" firstStartedPulling="2026-03-13 08:20:46.770550736 +0000 UTC m=+2506.441329728" lastFinishedPulling="2026-03-13 08:20:47.484109553 +0000 UTC m=+2507.154888555" observedRunningTime="2026-03-13 08:20:47.931993734 +0000 UTC m=+2507.602772726" watchObservedRunningTime="2026-03-13 08:20:47.938352501 +0000 UTC m=+2507.609131543" Mar 13 08:21:00 crc kubenswrapper[4876]: I0313 08:21:00.035930 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:21:00 crc kubenswrapper[4876]: E0313 08:21:00.036808 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:21:11 crc kubenswrapper[4876]: I0313 08:21:11.044518 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:21:11 crc kubenswrapper[4876]: E0313 08:21:11.045960 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:21:23 crc kubenswrapper[4876]: I0313 08:21:23.037821 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:21:23 crc kubenswrapper[4876]: E0313 08:21:23.038863 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:21:36 crc kubenswrapper[4876]: I0313 08:21:36.035693 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:21:36 crc kubenswrapper[4876]: E0313 08:21:36.036490 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:21:48 crc kubenswrapper[4876]: I0313 08:21:48.036113 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:21:48 crc kubenswrapper[4876]: E0313 08:21:48.036760 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.156413 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556502-pq9vk"] Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.158462 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.160863 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.161401 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.163492 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.180117 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556502-pq9vk"] Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.289469 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtcjj\" (UniqueName: \"kubernetes.io/projected/e0110e2c-4de2-4bc8-bf9a-fc8745032724-kube-api-access-qtcjj\") pod \"auto-csr-approver-29556502-pq9vk\" (UID: \"e0110e2c-4de2-4bc8-bf9a-fc8745032724\") " pod="openshift-infra/auto-csr-approver-29556502-pq9vk" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.393323 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtcjj\" (UniqueName: \"kubernetes.io/projected/e0110e2c-4de2-4bc8-bf9a-fc8745032724-kube-api-access-qtcjj\") pod \"auto-csr-approver-29556502-pq9vk\" (UID: \"e0110e2c-4de2-4bc8-bf9a-fc8745032724\") " pod="openshift-infra/auto-csr-approver-29556502-pq9vk" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.430402 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtcjj\" (UniqueName: \"kubernetes.io/projected/e0110e2c-4de2-4bc8-bf9a-fc8745032724-kube-api-access-qtcjj\") pod \"auto-csr-approver-29556502-pq9vk\" (UID: \"e0110e2c-4de2-4bc8-bf9a-fc8745032724\") " pod="openshift-infra/auto-csr-approver-29556502-pq9vk" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.483587 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" Mar 13 08:22:00 crc kubenswrapper[4876]: I0313 08:22:00.965054 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556502-pq9vk"] Mar 13 08:22:01 crc kubenswrapper[4876]: I0313 08:22:01.632575 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" event={"ID":"e0110e2c-4de2-4bc8-bf9a-fc8745032724","Type":"ContainerStarted","Data":"9ac335b0c10fe3eaea7fa167b89c4cead680ebc01be8d10b12cbc85a18885dbf"} Mar 13 08:22:02 crc kubenswrapper[4876]: I0313 08:22:02.036726 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:22:02 crc kubenswrapper[4876]: E0313 08:22:02.037227 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:22:02 crc kubenswrapper[4876]: I0313 08:22:02.643343 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" event={"ID":"e0110e2c-4de2-4bc8-bf9a-fc8745032724","Type":"ContainerStarted","Data":"ef1f57654297a7a7b06512607d8849db7960efc9e3e058ac36e25ea37154b4b5"} Mar 13 08:22:02 crc kubenswrapper[4876]: I0313 08:22:02.677363 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" podStartSLOduration=1.635340695 podStartE2EDuration="2.677329587s" podCreationTimestamp="2026-03-13 08:22:00 +0000 UTC" firstStartedPulling="2026-03-13 08:22:00.971798532 +0000 UTC m=+2580.642577514" lastFinishedPulling="2026-03-13 08:22:02.013787424 +0000 UTC m=+2581.684566406" observedRunningTime="2026-03-13 08:22:02.665135879 +0000 UTC m=+2582.335914881" watchObservedRunningTime="2026-03-13 08:22:02.677329587 +0000 UTC m=+2582.348108579" Mar 13 08:22:03 crc kubenswrapper[4876]: I0313 08:22:03.652574 4876 generic.go:334] "Generic (PLEG): container finished" podID="e0110e2c-4de2-4bc8-bf9a-fc8745032724" containerID="ef1f57654297a7a7b06512607d8849db7960efc9e3e058ac36e25ea37154b4b5" exitCode=0 Mar 13 08:22:03 crc kubenswrapper[4876]: I0313 08:22:03.652645 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" event={"ID":"e0110e2c-4de2-4bc8-bf9a-fc8745032724","Type":"ContainerDied","Data":"ef1f57654297a7a7b06512607d8849db7960efc9e3e058ac36e25ea37154b4b5"} Mar 13 08:22:04 crc kubenswrapper[4876]: I0313 08:22:04.988497 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" Mar 13 08:22:05 crc kubenswrapper[4876]: I0313 08:22:05.095901 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtcjj\" (UniqueName: \"kubernetes.io/projected/e0110e2c-4de2-4bc8-bf9a-fc8745032724-kube-api-access-qtcjj\") pod \"e0110e2c-4de2-4bc8-bf9a-fc8745032724\" (UID: \"e0110e2c-4de2-4bc8-bf9a-fc8745032724\") " Mar 13 08:22:05 crc kubenswrapper[4876]: I0313 08:22:05.103276 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0110e2c-4de2-4bc8-bf9a-fc8745032724-kube-api-access-qtcjj" (OuterVolumeSpecName: "kube-api-access-qtcjj") pod "e0110e2c-4de2-4bc8-bf9a-fc8745032724" (UID: "e0110e2c-4de2-4bc8-bf9a-fc8745032724"). InnerVolumeSpecName "kube-api-access-qtcjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:22:05 crc kubenswrapper[4876]: I0313 08:22:05.197960 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtcjj\" (UniqueName: \"kubernetes.io/projected/e0110e2c-4de2-4bc8-bf9a-fc8745032724-kube-api-access-qtcjj\") on node \"crc\" DevicePath \"\"" Mar 13 08:22:05 crc kubenswrapper[4876]: I0313 08:22:05.670476 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" event={"ID":"e0110e2c-4de2-4bc8-bf9a-fc8745032724","Type":"ContainerDied","Data":"9ac335b0c10fe3eaea7fa167b89c4cead680ebc01be8d10b12cbc85a18885dbf"} Mar 13 08:22:05 crc kubenswrapper[4876]: I0313 08:22:05.670517 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ac335b0c10fe3eaea7fa167b89c4cead680ebc01be8d10b12cbc85a18885dbf" Mar 13 08:22:05 crc kubenswrapper[4876]: I0313 08:22:05.670529 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556502-pq9vk" Mar 13 08:22:06 crc kubenswrapper[4876]: I0313 08:22:06.065324 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556496-xsw4v"] Mar 13 08:22:06 crc kubenswrapper[4876]: I0313 08:22:06.073543 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556496-xsw4v"] Mar 13 08:22:07 crc kubenswrapper[4876]: I0313 08:22:07.046096 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35b8ae2c-4713-4cab-8118-0f5ef2df30d5" path="/var/lib/kubelet/pods/35b8ae2c-4713-4cab-8118-0f5ef2df30d5/volumes" Mar 13 08:22:16 crc kubenswrapper[4876]: I0313 08:22:16.035968 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:22:16 crc kubenswrapper[4876]: E0313 08:22:16.037342 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:22:23 crc kubenswrapper[4876]: I0313 08:22:23.348855 4876 scope.go:117] "RemoveContainer" containerID="5a437662c3ad3512f2c539e65fc43f38b07d2770206c2545cda4d07df85a0d1a" Mar 13 08:22:31 crc kubenswrapper[4876]: I0313 08:22:31.043532 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:22:31 crc kubenswrapper[4876]: I0313 08:22:31.942493 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"87c133791187037b4f20020f848722bad2719b25256974c3c1a18083148ed7f2"} Mar 13 08:23:11 crc kubenswrapper[4876]: I0313 08:23:11.311038 4876 generic.go:334] "Generic (PLEG): container finished" podID="18be12a3-56e2-4b39-8c8f-784a7cafed2b" containerID="771e73610b3fd6e6da486126216413fc946ed3fd7e26b83b11a5a17cca9cf14e" exitCode=0 Mar 13 08:23:11 crc kubenswrapper[4876]: I0313 08:23:11.311135 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" event={"ID":"18be12a3-56e2-4b39-8c8f-784a7cafed2b","Type":"ContainerDied","Data":"771e73610b3fd6e6da486126216413fc946ed3fd7e26b83b11a5a17cca9cf14e"} Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.727758 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.846794 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ssh-key-openstack-edpm-ipam\") pod \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.847300 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-inventory\") pod \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.847449 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-2\") pod \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.847545 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-telemetry-combined-ca-bundle\") pod \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.847648 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-0\") pod \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.847776 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plcpn\" (UniqueName: \"kubernetes.io/projected/18be12a3-56e2-4b39-8c8f-784a7cafed2b-kube-api-access-plcpn\") pod \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.847868 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-1\") pod \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\" (UID: \"18be12a3-56e2-4b39-8c8f-784a7cafed2b\") " Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.854401 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18be12a3-56e2-4b39-8c8f-784a7cafed2b-kube-api-access-plcpn" (OuterVolumeSpecName: "kube-api-access-plcpn") pod "18be12a3-56e2-4b39-8c8f-784a7cafed2b" (UID: "18be12a3-56e2-4b39-8c8f-784a7cafed2b"). InnerVolumeSpecName "kube-api-access-plcpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.854459 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "18be12a3-56e2-4b39-8c8f-784a7cafed2b" (UID: "18be12a3-56e2-4b39-8c8f-784a7cafed2b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.875843 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "18be12a3-56e2-4b39-8c8f-784a7cafed2b" (UID: "18be12a3-56e2-4b39-8c8f-784a7cafed2b"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.881546 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "18be12a3-56e2-4b39-8c8f-784a7cafed2b" (UID: "18be12a3-56e2-4b39-8c8f-784a7cafed2b"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.881733 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "18be12a3-56e2-4b39-8c8f-784a7cafed2b" (UID: "18be12a3-56e2-4b39-8c8f-784a7cafed2b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.884989 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-inventory" (OuterVolumeSpecName: "inventory") pod "18be12a3-56e2-4b39-8c8f-784a7cafed2b" (UID: "18be12a3-56e2-4b39-8c8f-784a7cafed2b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.891936 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "18be12a3-56e2-4b39-8c8f-784a7cafed2b" (UID: "18be12a3-56e2-4b39-8c8f-784a7cafed2b"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.950260 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.950295 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.950307 4876 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.950317 4876 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.950333 4876 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.950347 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plcpn\" (UniqueName: \"kubernetes.io/projected/18be12a3-56e2-4b39-8c8f-784a7cafed2b-kube-api-access-plcpn\") on node \"crc\" DevicePath \"\"" Mar 13 08:23:12 crc kubenswrapper[4876]: I0313 08:23:12.950360 4876 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/18be12a3-56e2-4b39-8c8f-784a7cafed2b-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 13 08:23:13 crc kubenswrapper[4876]: I0313 08:23:13.341048 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" event={"ID":"18be12a3-56e2-4b39-8c8f-784a7cafed2b","Type":"ContainerDied","Data":"08c80c63531bcd376f8616a1cfab25b4a0020e7ba8869d99643b5663a037a69a"} Mar 13 08:23:13 crc kubenswrapper[4876]: I0313 08:23:13.341740 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08c80c63531bcd376f8616a1cfab25b4a0020e7ba8869d99643b5663a037a69a" Mar 13 08:23:13 crc kubenswrapper[4876]: I0313 08:23:13.341510 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx" Mar 13 08:23:35 crc kubenswrapper[4876]: E0313 08:23:35.329430 4876 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.39:58496->38.102.83.39:46711: read tcp 38.102.83.39:58496->38.102.83.39:46711: read: connection reset by peer Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.776982 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Mar 13 08:23:53 crc kubenswrapper[4876]: E0313 08:23:53.778225 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0110e2c-4de2-4bc8-bf9a-fc8745032724" containerName="oc" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.778277 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0110e2c-4de2-4bc8-bf9a-fc8745032724" containerName="oc" Mar 13 08:23:53 crc kubenswrapper[4876]: E0313 08:23:53.778329 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18be12a3-56e2-4b39-8c8f-784a7cafed2b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.778342 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="18be12a3-56e2-4b39-8c8f-784a7cafed2b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.778672 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0110e2c-4de2-4bc8-bf9a-fc8745032724" containerName="oc" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.778702 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="18be12a3-56e2-4b39-8c8f-784a7cafed2b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.779643 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.781962 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.782044 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.782623 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.783228 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nrc9h" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.795210 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.954695 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.954821 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.954884 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpj5x\" (UniqueName: \"kubernetes.io/projected/009d8699-15dd-4c89-bd6c-5e7a936e1b59-kube-api-access-fpj5x\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.955186 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.955282 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.955381 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.955449 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.955554 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-config-data\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:53 crc kubenswrapper[4876]: I0313 08:23:53.955628 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.058481 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.059047 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-config-data\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.059164 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.059349 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.059492 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.059810 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpj5x\" (UniqueName: \"kubernetes.io/projected/009d8699-15dd-4c89-bd6c-5e7a936e1b59-kube-api-access-fpj5x\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.059978 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.060085 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.060199 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.060426 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.061044 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.061191 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.061474 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-config-data\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.061691 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.066068 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.066531 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.075148 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.079881 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpj5x\" (UniqueName: \"kubernetes.io/projected/009d8699-15dd-4c89-bd6c-5e7a936e1b59-kube-api-access-fpj5x\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.099700 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.108700 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.581481 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 13 08:23:54 crc kubenswrapper[4876]: I0313 08:23:54.785957 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"009d8699-15dd-4c89-bd6c-5e7a936e1b59","Type":"ContainerStarted","Data":"4c7a7a787683777f0ea7a5daf1786ae418fa67b1459680d04d8d03a9844dccaa"} Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.180102 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556504-2mrbk"] Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.182203 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556504-2mrbk" Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.185080 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.185120 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.185401 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.190615 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556504-2mrbk"] Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.316843 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f6k6\" (UniqueName: \"kubernetes.io/projected/0eba926b-7bb7-4522-bc87-94f5ea0ad6f4-kube-api-access-8f6k6\") pod \"auto-csr-approver-29556504-2mrbk\" (UID: \"0eba926b-7bb7-4522-bc87-94f5ea0ad6f4\") " pod="openshift-infra/auto-csr-approver-29556504-2mrbk" Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.419589 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f6k6\" (UniqueName: \"kubernetes.io/projected/0eba926b-7bb7-4522-bc87-94f5ea0ad6f4-kube-api-access-8f6k6\") pod \"auto-csr-approver-29556504-2mrbk\" (UID: \"0eba926b-7bb7-4522-bc87-94f5ea0ad6f4\") " pod="openshift-infra/auto-csr-approver-29556504-2mrbk" Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.448433 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f6k6\" (UniqueName: \"kubernetes.io/projected/0eba926b-7bb7-4522-bc87-94f5ea0ad6f4-kube-api-access-8f6k6\") pod \"auto-csr-approver-29556504-2mrbk\" (UID: \"0eba926b-7bb7-4522-bc87-94f5ea0ad6f4\") " pod="openshift-infra/auto-csr-approver-29556504-2mrbk" Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.511066 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556504-2mrbk" Mar 13 08:24:00 crc kubenswrapper[4876]: I0313 08:24:00.947968 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556504-2mrbk"] Mar 13 08:24:01 crc kubenswrapper[4876]: I0313 08:24:01.874582 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556504-2mrbk" event={"ID":"0eba926b-7bb7-4522-bc87-94f5ea0ad6f4","Type":"ContainerStarted","Data":"e46aa9b6bf116285ef6b3e4010770393148049d1ec3f7042f6239e5cf7d78a2a"} Mar 13 08:24:02 crc kubenswrapper[4876]: I0313 08:24:02.883972 4876 generic.go:334] "Generic (PLEG): container finished" podID="0eba926b-7bb7-4522-bc87-94f5ea0ad6f4" containerID="8cbcd17c6bff74394c479fba13f47c4e28fb41c91324bde5455471f54d1a7053" exitCode=0 Mar 13 08:24:02 crc kubenswrapper[4876]: I0313 08:24:02.884031 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556504-2mrbk" event={"ID":"0eba926b-7bb7-4522-bc87-94f5ea0ad6f4","Type":"ContainerDied","Data":"8cbcd17c6bff74394c479fba13f47c4e28fb41c91324bde5455471f54d1a7053"} Mar 13 08:24:04 crc kubenswrapper[4876]: I0313 08:24:04.247041 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556504-2mrbk" Mar 13 08:24:04 crc kubenswrapper[4876]: I0313 08:24:04.297312 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f6k6\" (UniqueName: \"kubernetes.io/projected/0eba926b-7bb7-4522-bc87-94f5ea0ad6f4-kube-api-access-8f6k6\") pod \"0eba926b-7bb7-4522-bc87-94f5ea0ad6f4\" (UID: \"0eba926b-7bb7-4522-bc87-94f5ea0ad6f4\") " Mar 13 08:24:04 crc kubenswrapper[4876]: I0313 08:24:04.303603 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eba926b-7bb7-4522-bc87-94f5ea0ad6f4-kube-api-access-8f6k6" (OuterVolumeSpecName: "kube-api-access-8f6k6") pod "0eba926b-7bb7-4522-bc87-94f5ea0ad6f4" (UID: "0eba926b-7bb7-4522-bc87-94f5ea0ad6f4"). InnerVolumeSpecName "kube-api-access-8f6k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:24:04 crc kubenswrapper[4876]: I0313 08:24:04.399465 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f6k6\" (UniqueName: \"kubernetes.io/projected/0eba926b-7bb7-4522-bc87-94f5ea0ad6f4-kube-api-access-8f6k6\") on node \"crc\" DevicePath \"\"" Mar 13 08:24:04 crc kubenswrapper[4876]: I0313 08:24:04.905930 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556504-2mrbk" event={"ID":"0eba926b-7bb7-4522-bc87-94f5ea0ad6f4","Type":"ContainerDied","Data":"e46aa9b6bf116285ef6b3e4010770393148049d1ec3f7042f6239e5cf7d78a2a"} Mar 13 08:24:04 crc kubenswrapper[4876]: I0313 08:24:04.905968 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e46aa9b6bf116285ef6b3e4010770393148049d1ec3f7042f6239e5cf7d78a2a" Mar 13 08:24:04 crc kubenswrapper[4876]: I0313 08:24:04.905971 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556504-2mrbk" Mar 13 08:24:05 crc kubenswrapper[4876]: I0313 08:24:05.330030 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556498-mv7bp"] Mar 13 08:24:05 crc kubenswrapper[4876]: I0313 08:24:05.340333 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556498-mv7bp"] Mar 13 08:24:07 crc kubenswrapper[4876]: I0313 08:24:07.049764 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5" path="/var/lib/kubelet/pods/80b34b33-f2c3-4eb7-9566-2a7c39a3e6a5/volumes" Mar 13 08:24:20 crc kubenswrapper[4876]: E0313 08:24:20.861840 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Mar 13 08:24:20 crc kubenswrapper[4876]: E0313 08:24:20.864888 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fpj5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(009d8699-15dd-4c89-bd6c-5e7a936e1b59): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 08:24:20 crc kubenswrapper[4876]: E0313 08:24:20.866190 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="009d8699-15dd-4c89-bd6c-5e7a936e1b59" Mar 13 08:24:21 crc kubenswrapper[4876]: E0313 08:24:21.082814 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="009d8699-15dd-4c89-bd6c-5e7a936e1b59" Mar 13 08:24:23 crc kubenswrapper[4876]: I0313 08:24:23.473179 4876 scope.go:117] "RemoveContainer" containerID="6bfcf56cb337dd671211118065de14c3cddbe0810628f832b62b06a1648733ed" Mar 13 08:24:32 crc kubenswrapper[4876]: I0313 08:24:32.484785 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 13 08:24:34 crc kubenswrapper[4876]: I0313 08:24:34.238573 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"009d8699-15dd-4c89-bd6c-5e7a936e1b59","Type":"ContainerStarted","Data":"583433128f332739101afda35726f7f69c6d0148f2ad7e4a97c110717ce4ce3a"} Mar 13 08:24:34 crc kubenswrapper[4876]: I0313 08:24:34.265928 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.376320306 podStartE2EDuration="42.265910582s" podCreationTimestamp="2026-03-13 08:23:52 +0000 UTC" firstStartedPulling="2026-03-13 08:23:54.591567988 +0000 UTC m=+2694.262346970" lastFinishedPulling="2026-03-13 08:24:32.481158234 +0000 UTC m=+2732.151937246" observedRunningTime="2026-03-13 08:24:34.25789903 +0000 UTC m=+2733.928678012" watchObservedRunningTime="2026-03-13 08:24:34.265910582 +0000 UTC m=+2733.936689564" Mar 13 08:24:54 crc kubenswrapper[4876]: I0313 08:24:54.607141 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:24:54 crc kubenswrapper[4876]: I0313 08:24:54.607651 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.280798 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z659d"] Mar 13 08:24:55 crc kubenswrapper[4876]: E0313 08:24:55.282051 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eba926b-7bb7-4522-bc87-94f5ea0ad6f4" containerName="oc" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.282075 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eba926b-7bb7-4522-bc87-94f5ea0ad6f4" containerName="oc" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.282337 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eba926b-7bb7-4522-bc87-94f5ea0ad6f4" containerName="oc" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.283700 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.307154 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z659d"] Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.317325 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-catalog-content\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.317606 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-utilities\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.317760 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbd4w\" (UniqueName: \"kubernetes.io/projected/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-kube-api-access-sbd4w\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.419040 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbd4w\" (UniqueName: \"kubernetes.io/projected/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-kube-api-access-sbd4w\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.419121 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-catalog-content\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.419176 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-utilities\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.419694 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-utilities\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.419799 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-catalog-content\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.445088 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbd4w\" (UniqueName: \"kubernetes.io/projected/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-kube-api-access-sbd4w\") pod \"redhat-operators-z659d\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:55 crc kubenswrapper[4876]: I0313 08:24:55.611428 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:24:56 crc kubenswrapper[4876]: I0313 08:24:56.029579 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z659d"] Mar 13 08:24:56 crc kubenswrapper[4876]: W0313 08:24:56.036223 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod901fd93c_f641_4e0c_ae37_5133e9c4ea5c.slice/crio-79f2138c9d1d2705da85870876dbd6b21fd60009ccedea481f7f2750c9cc3b35 WatchSource:0}: Error finding container 79f2138c9d1d2705da85870876dbd6b21fd60009ccedea481f7f2750c9cc3b35: Status 404 returned error can't find the container with id 79f2138c9d1d2705da85870876dbd6b21fd60009ccedea481f7f2750c9cc3b35 Mar 13 08:24:56 crc kubenswrapper[4876]: I0313 08:24:56.462378 4876 generic.go:334] "Generic (PLEG): container finished" podID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerID="cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316" exitCode=0 Mar 13 08:24:56 crc kubenswrapper[4876]: I0313 08:24:56.462425 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z659d" event={"ID":"901fd93c-f641-4e0c-ae37-5133e9c4ea5c","Type":"ContainerDied","Data":"cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316"} Mar 13 08:24:56 crc kubenswrapper[4876]: I0313 08:24:56.462450 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z659d" event={"ID":"901fd93c-f641-4e0c-ae37-5133e9c4ea5c","Type":"ContainerStarted","Data":"79f2138c9d1d2705da85870876dbd6b21fd60009ccedea481f7f2750c9cc3b35"} Mar 13 08:24:57 crc kubenswrapper[4876]: I0313 08:24:57.476053 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z659d" event={"ID":"901fd93c-f641-4e0c-ae37-5133e9c4ea5c","Type":"ContainerStarted","Data":"083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439"} Mar 13 08:25:00 crc kubenswrapper[4876]: I0313 08:25:00.509205 4876 generic.go:334] "Generic (PLEG): container finished" podID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerID="083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439" exitCode=0 Mar 13 08:25:00 crc kubenswrapper[4876]: I0313 08:25:00.509283 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z659d" event={"ID":"901fd93c-f641-4e0c-ae37-5133e9c4ea5c","Type":"ContainerDied","Data":"083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439"} Mar 13 08:25:01 crc kubenswrapper[4876]: I0313 08:25:01.521852 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z659d" event={"ID":"901fd93c-f641-4e0c-ae37-5133e9c4ea5c","Type":"ContainerStarted","Data":"f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81"} Mar 13 08:25:01 crc kubenswrapper[4876]: I0313 08:25:01.541460 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z659d" podStartSLOduration=2.105319088 podStartE2EDuration="6.541440812s" podCreationTimestamp="2026-03-13 08:24:55 +0000 UTC" firstStartedPulling="2026-03-13 08:24:56.464342899 +0000 UTC m=+2756.135121881" lastFinishedPulling="2026-03-13 08:25:00.900464623 +0000 UTC m=+2760.571243605" observedRunningTime="2026-03-13 08:25:01.538468319 +0000 UTC m=+2761.209247301" watchObservedRunningTime="2026-03-13 08:25:01.541440812 +0000 UTC m=+2761.212219794" Mar 13 08:25:05 crc kubenswrapper[4876]: I0313 08:25:05.613191 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:25:05 crc kubenswrapper[4876]: I0313 08:25:05.614176 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:25:06 crc kubenswrapper[4876]: I0313 08:25:06.659949 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z659d" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="registry-server" probeResult="failure" output=< Mar 13 08:25:06 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:25:06 crc kubenswrapper[4876]: > Mar 13 08:25:15 crc kubenswrapper[4876]: I0313 08:25:15.665569 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:25:15 crc kubenswrapper[4876]: I0313 08:25:15.725093 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:25:15 crc kubenswrapper[4876]: I0313 08:25:15.914539 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z659d"] Mar 13 08:25:16 crc kubenswrapper[4876]: I0313 08:25:16.712371 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z659d" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="registry-server" containerID="cri-o://f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81" gracePeriod=2 Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.175015 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.368146 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbd4w\" (UniqueName: \"kubernetes.io/projected/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-kube-api-access-sbd4w\") pod \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.368194 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-utilities\") pod \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.368376 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-catalog-content\") pod \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\" (UID: \"901fd93c-f641-4e0c-ae37-5133e9c4ea5c\") " Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.369133 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-utilities" (OuterVolumeSpecName: "utilities") pod "901fd93c-f641-4e0c-ae37-5133e9c4ea5c" (UID: "901fd93c-f641-4e0c-ae37-5133e9c4ea5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.386455 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-kube-api-access-sbd4w" (OuterVolumeSpecName: "kube-api-access-sbd4w") pod "901fd93c-f641-4e0c-ae37-5133e9c4ea5c" (UID: "901fd93c-f641-4e0c-ae37-5133e9c4ea5c"). InnerVolumeSpecName "kube-api-access-sbd4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.471303 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbd4w\" (UniqueName: \"kubernetes.io/projected/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-kube-api-access-sbd4w\") on node \"crc\" DevicePath \"\"" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.471356 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.508938 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "901fd93c-f641-4e0c-ae37-5133e9c4ea5c" (UID: "901fd93c-f641-4e0c-ae37-5133e9c4ea5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.573849 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901fd93c-f641-4e0c-ae37-5133e9c4ea5c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.726864 4876 generic.go:334] "Generic (PLEG): container finished" podID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerID="f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81" exitCode=0 Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.726913 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z659d" event={"ID":"901fd93c-f641-4e0c-ae37-5133e9c4ea5c","Type":"ContainerDied","Data":"f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81"} Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.726951 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z659d" event={"ID":"901fd93c-f641-4e0c-ae37-5133e9c4ea5c","Type":"ContainerDied","Data":"79f2138c9d1d2705da85870876dbd6b21fd60009ccedea481f7f2750c9cc3b35"} Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.726971 4876 scope.go:117] "RemoveContainer" containerID="f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.726986 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z659d" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.752323 4876 scope.go:117] "RemoveContainer" containerID="083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.778963 4876 scope.go:117] "RemoveContainer" containerID="cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.837977 4876 scope.go:117] "RemoveContainer" containerID="f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81" Mar 13 08:25:17 crc kubenswrapper[4876]: E0313 08:25:17.838736 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81\": container with ID starting with f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81 not found: ID does not exist" containerID="f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.838773 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81"} err="failed to get container status \"f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81\": rpc error: code = NotFound desc = could not find container \"f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81\": container with ID starting with f84ad3f465185aacb31109a9ba4fa78213b141dfdc30a5fd3e35572933962b81 not found: ID does not exist" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.838825 4876 scope.go:117] "RemoveContainer" containerID="083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439" Mar 13 08:25:17 crc kubenswrapper[4876]: E0313 08:25:17.839210 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439\": container with ID starting with 083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439 not found: ID does not exist" containerID="083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.839282 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439"} err="failed to get container status \"083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439\": rpc error: code = NotFound desc = could not find container \"083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439\": container with ID starting with 083aadd08335215150aca5e0b04bec90f9ced7b05b28fc46a7236af27a606439 not found: ID does not exist" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.839316 4876 scope.go:117] "RemoveContainer" containerID="cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316" Mar 13 08:25:17 crc kubenswrapper[4876]: E0313 08:25:17.839634 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316\": container with ID starting with cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316 not found: ID does not exist" containerID="cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.839664 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316"} err="failed to get container status \"cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316\": rpc error: code = NotFound desc = could not find container \"cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316\": container with ID starting with cab74dd54f662de70b84d0a03b25e22168fc48c6e167066923615e22310b5316 not found: ID does not exist" Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.843489 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z659d"] Mar 13 08:25:17 crc kubenswrapper[4876]: I0313 08:25:17.856228 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z659d"] Mar 13 08:25:19 crc kubenswrapper[4876]: I0313 08:25:19.046911 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" path="/var/lib/kubelet/pods/901fd93c-f641-4e0c-ae37-5133e9c4ea5c/volumes" Mar 13 08:25:24 crc kubenswrapper[4876]: I0313 08:25:24.608410 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:25:24 crc kubenswrapper[4876]: I0313 08:25:24.611071 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:25:54 crc kubenswrapper[4876]: I0313 08:25:54.607955 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:25:54 crc kubenswrapper[4876]: I0313 08:25:54.608755 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:25:54 crc kubenswrapper[4876]: I0313 08:25:54.608828 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:25:54 crc kubenswrapper[4876]: I0313 08:25:54.609869 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"87c133791187037b4f20020f848722bad2719b25256974c3c1a18083148ed7f2"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:25:54 crc kubenswrapper[4876]: I0313 08:25:54.609971 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://87c133791187037b4f20020f848722bad2719b25256974c3c1a18083148ed7f2" gracePeriod=600 Mar 13 08:25:55 crc kubenswrapper[4876]: I0313 08:25:55.100308 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="87c133791187037b4f20020f848722bad2719b25256974c3c1a18083148ed7f2" exitCode=0 Mar 13 08:25:55 crc kubenswrapper[4876]: I0313 08:25:55.100367 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"87c133791187037b4f20020f848722bad2719b25256974c3c1a18083148ed7f2"} Mar 13 08:25:55 crc kubenswrapper[4876]: I0313 08:25:55.100772 4876 scope.go:117] "RemoveContainer" containerID="b5747c8fd42c76b37db212d2dd0869248aebbc01ffa390b16a4b8ced708ccb66" Mar 13 08:25:55 crc kubenswrapper[4876]: I0313 08:25:55.100650 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b"} Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.166281 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556506-z26rs"] Mar 13 08:26:00 crc kubenswrapper[4876]: E0313 08:26:00.168447 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="registry-server" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.168559 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="registry-server" Mar 13 08:26:00 crc kubenswrapper[4876]: E0313 08:26:00.168659 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="extract-utilities" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.168759 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="extract-utilities" Mar 13 08:26:00 crc kubenswrapper[4876]: E0313 08:26:00.168850 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="extract-content" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.168937 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="extract-content" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.169267 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="901fd93c-f641-4e0c-ae37-5133e9c4ea5c" containerName="registry-server" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.170131 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556506-z26rs" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.176800 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.176902 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.177295 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.177983 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556506-z26rs"] Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.222534 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7pzq\" (UniqueName: \"kubernetes.io/projected/ff7c968c-dac3-4d92-bd12-7334775430c7-kube-api-access-z7pzq\") pod \"auto-csr-approver-29556506-z26rs\" (UID: \"ff7c968c-dac3-4d92-bd12-7334775430c7\") " pod="openshift-infra/auto-csr-approver-29556506-z26rs" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.324222 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7pzq\" (UniqueName: \"kubernetes.io/projected/ff7c968c-dac3-4d92-bd12-7334775430c7-kube-api-access-z7pzq\") pod \"auto-csr-approver-29556506-z26rs\" (UID: \"ff7c968c-dac3-4d92-bd12-7334775430c7\") " pod="openshift-infra/auto-csr-approver-29556506-z26rs" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.342169 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7pzq\" (UniqueName: \"kubernetes.io/projected/ff7c968c-dac3-4d92-bd12-7334775430c7-kube-api-access-z7pzq\") pod \"auto-csr-approver-29556506-z26rs\" (UID: \"ff7c968c-dac3-4d92-bd12-7334775430c7\") " pod="openshift-infra/auto-csr-approver-29556506-z26rs" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.493630 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556506-z26rs" Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.929107 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556506-z26rs"] Mar 13 08:26:00 crc kubenswrapper[4876]: I0313 08:26:00.934657 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:26:01 crc kubenswrapper[4876]: I0313 08:26:01.168478 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556506-z26rs" event={"ID":"ff7c968c-dac3-4d92-bd12-7334775430c7","Type":"ContainerStarted","Data":"46c47a1037daa8414d90039c6315c645ac7b21117e5f6e8f713e21bd9a684040"} Mar 13 08:26:02 crc kubenswrapper[4876]: I0313 08:26:02.181946 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556506-z26rs" event={"ID":"ff7c968c-dac3-4d92-bd12-7334775430c7","Type":"ContainerStarted","Data":"f359a0e13dcb77d3e97aa0d981d74018b291a3aaf838877976fca83c5fc98f94"} Mar 13 08:26:02 crc kubenswrapper[4876]: I0313 08:26:02.200556 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556506-z26rs" podStartSLOduration=1.434583925 podStartE2EDuration="2.200537224s" podCreationTimestamp="2026-03-13 08:26:00 +0000 UTC" firstStartedPulling="2026-03-13 08:26:00.934414446 +0000 UTC m=+2820.605193428" lastFinishedPulling="2026-03-13 08:26:01.700367745 +0000 UTC m=+2821.371146727" observedRunningTime="2026-03-13 08:26:02.193513939 +0000 UTC m=+2821.864292921" watchObservedRunningTime="2026-03-13 08:26:02.200537224 +0000 UTC m=+2821.871316206" Mar 13 08:26:03 crc kubenswrapper[4876]: I0313 08:26:03.196442 4876 generic.go:334] "Generic (PLEG): container finished" podID="ff7c968c-dac3-4d92-bd12-7334775430c7" containerID="f359a0e13dcb77d3e97aa0d981d74018b291a3aaf838877976fca83c5fc98f94" exitCode=0 Mar 13 08:26:03 crc kubenswrapper[4876]: I0313 08:26:03.196730 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556506-z26rs" event={"ID":"ff7c968c-dac3-4d92-bd12-7334775430c7","Type":"ContainerDied","Data":"f359a0e13dcb77d3e97aa0d981d74018b291a3aaf838877976fca83c5fc98f94"} Mar 13 08:26:04 crc kubenswrapper[4876]: I0313 08:26:04.605669 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556506-z26rs" Mar 13 08:26:04 crc kubenswrapper[4876]: I0313 08:26:04.724525 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7pzq\" (UniqueName: \"kubernetes.io/projected/ff7c968c-dac3-4d92-bd12-7334775430c7-kube-api-access-z7pzq\") pod \"ff7c968c-dac3-4d92-bd12-7334775430c7\" (UID: \"ff7c968c-dac3-4d92-bd12-7334775430c7\") " Mar 13 08:26:04 crc kubenswrapper[4876]: I0313 08:26:04.732444 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7c968c-dac3-4d92-bd12-7334775430c7-kube-api-access-z7pzq" (OuterVolumeSpecName: "kube-api-access-z7pzq") pod "ff7c968c-dac3-4d92-bd12-7334775430c7" (UID: "ff7c968c-dac3-4d92-bd12-7334775430c7"). InnerVolumeSpecName "kube-api-access-z7pzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:26:04 crc kubenswrapper[4876]: I0313 08:26:04.827595 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7pzq\" (UniqueName: \"kubernetes.io/projected/ff7c968c-dac3-4d92-bd12-7334775430c7-kube-api-access-z7pzq\") on node \"crc\" DevicePath \"\"" Mar 13 08:26:05 crc kubenswrapper[4876]: I0313 08:26:05.215897 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556506-z26rs" event={"ID":"ff7c968c-dac3-4d92-bd12-7334775430c7","Type":"ContainerDied","Data":"46c47a1037daa8414d90039c6315c645ac7b21117e5f6e8f713e21bd9a684040"} Mar 13 08:26:05 crc kubenswrapper[4876]: I0313 08:26:05.216285 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46c47a1037daa8414d90039c6315c645ac7b21117e5f6e8f713e21bd9a684040" Mar 13 08:26:05 crc kubenswrapper[4876]: I0313 08:26:05.215951 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556506-z26rs" Mar 13 08:26:05 crc kubenswrapper[4876]: I0313 08:26:05.266391 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556500-fg284"] Mar 13 08:26:05 crc kubenswrapper[4876]: I0313 08:26:05.274659 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556500-fg284"] Mar 13 08:26:07 crc kubenswrapper[4876]: I0313 08:26:07.049509 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5024f93c-6cfc-4019-9ad4-b39aee51d02e" path="/var/lib/kubelet/pods/5024f93c-6cfc-4019-9ad4-b39aee51d02e/volumes" Mar 13 08:26:23 crc kubenswrapper[4876]: I0313 08:26:23.583033 4876 scope.go:117] "RemoveContainer" containerID="7745192dd8ac3dab7eac784d445a7a890627ec3e88601669549e11c6fe1d0a07" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.684284 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qtrmp"] Mar 13 08:26:54 crc kubenswrapper[4876]: E0313 08:26:54.686893 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7c968c-dac3-4d92-bd12-7334775430c7" containerName="oc" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.686923 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7c968c-dac3-4d92-bd12-7334775430c7" containerName="oc" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.687240 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7c968c-dac3-4d92-bd12-7334775430c7" containerName="oc" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.688948 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.711125 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtrmp"] Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.788043 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-utilities\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.788154 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjqq9\" (UniqueName: \"kubernetes.io/projected/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-kube-api-access-vjqq9\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.788204 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-catalog-content\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.889491 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-utilities\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.889550 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjqq9\" (UniqueName: \"kubernetes.io/projected/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-kube-api-access-vjqq9\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.889574 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-catalog-content\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.889941 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-utilities\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.890009 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-catalog-content\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:54 crc kubenswrapper[4876]: I0313 08:26:54.923209 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjqq9\" (UniqueName: \"kubernetes.io/projected/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-kube-api-access-vjqq9\") pod \"redhat-marketplace-qtrmp\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:55 crc kubenswrapper[4876]: I0313 08:26:55.017181 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:26:55 crc kubenswrapper[4876]: I0313 08:26:55.500641 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtrmp"] Mar 13 08:26:55 crc kubenswrapper[4876]: I0313 08:26:55.851641 4876 generic.go:334] "Generic (PLEG): container finished" podID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerID="d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e" exitCode=0 Mar 13 08:26:55 crc kubenswrapper[4876]: I0313 08:26:55.851703 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtrmp" event={"ID":"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd","Type":"ContainerDied","Data":"d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e"} Mar 13 08:26:55 crc kubenswrapper[4876]: I0313 08:26:55.851731 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtrmp" event={"ID":"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd","Type":"ContainerStarted","Data":"0eb2451c0d6a32eb1cbc62d2ae0c04694821c1d61d2b5de89406e62ab8c2007c"} Mar 13 08:26:56 crc kubenswrapper[4876]: I0313 08:26:56.862937 4876 generic.go:334] "Generic (PLEG): container finished" podID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerID="6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8" exitCode=0 Mar 13 08:26:56 crc kubenswrapper[4876]: I0313 08:26:56.863007 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtrmp" event={"ID":"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd","Type":"ContainerDied","Data":"6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8"} Mar 13 08:26:57 crc kubenswrapper[4876]: I0313 08:26:57.876986 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtrmp" event={"ID":"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd","Type":"ContainerStarted","Data":"d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2"} Mar 13 08:26:57 crc kubenswrapper[4876]: I0313 08:26:57.903689 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qtrmp" podStartSLOduration=2.482179375 podStartE2EDuration="3.903663124s" podCreationTimestamp="2026-03-13 08:26:54 +0000 UTC" firstStartedPulling="2026-03-13 08:26:55.853091366 +0000 UTC m=+2875.523870348" lastFinishedPulling="2026-03-13 08:26:57.274575105 +0000 UTC m=+2876.945354097" observedRunningTime="2026-03-13 08:26:57.893684348 +0000 UTC m=+2877.564463330" watchObservedRunningTime="2026-03-13 08:26:57.903663124 +0000 UTC m=+2877.574442106" Mar 13 08:27:05 crc kubenswrapper[4876]: I0313 08:27:05.018279 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:27:05 crc kubenswrapper[4876]: I0313 08:27:05.018921 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:27:05 crc kubenswrapper[4876]: I0313 08:27:05.079431 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:27:06 crc kubenswrapper[4876]: I0313 08:27:06.033737 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:27:06 crc kubenswrapper[4876]: I0313 08:27:06.093024 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtrmp"] Mar 13 08:27:07 crc kubenswrapper[4876]: I0313 08:27:07.988108 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qtrmp" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerName="registry-server" containerID="cri-o://d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2" gracePeriod=2 Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.507088 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.663726 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-catalog-content\") pod \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.663840 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjqq9\" (UniqueName: \"kubernetes.io/projected/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-kube-api-access-vjqq9\") pod \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.664051 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-utilities\") pod \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\" (UID: \"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd\") " Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.664988 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-utilities" (OuterVolumeSpecName: "utilities") pod "1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" (UID: "1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.677068 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-kube-api-access-vjqq9" (OuterVolumeSpecName: "kube-api-access-vjqq9") pod "1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" (UID: "1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd"). InnerVolumeSpecName "kube-api-access-vjqq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.686952 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" (UID: "1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.766844 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.766902 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.766925 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjqq9\" (UniqueName: \"kubernetes.io/projected/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd-kube-api-access-vjqq9\") on node \"crc\" DevicePath \"\"" Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.998315 4876 generic.go:334] "Generic (PLEG): container finished" podID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerID="d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2" exitCode=0 Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.998359 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtrmp" event={"ID":"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd","Type":"ContainerDied","Data":"d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2"} Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.998388 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtrmp" event={"ID":"1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd","Type":"ContainerDied","Data":"0eb2451c0d6a32eb1cbc62d2ae0c04694821c1d61d2b5de89406e62ab8c2007c"} Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.998406 4876 scope.go:117] "RemoveContainer" containerID="d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2" Mar 13 08:27:08 crc kubenswrapper[4876]: I0313 08:27:08.998406 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtrmp" Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.015791 4876 scope.go:117] "RemoveContainer" containerID="6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8" Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.049178 4876 scope.go:117] "RemoveContainer" containerID="d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e" Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.054180 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtrmp"] Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.054226 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtrmp"] Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.082205 4876 scope.go:117] "RemoveContainer" containerID="d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2" Mar 13 08:27:09 crc kubenswrapper[4876]: E0313 08:27:09.083516 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2\": container with ID starting with d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2 not found: ID does not exist" containerID="d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2" Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.083637 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2"} err="failed to get container status \"d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2\": rpc error: code = NotFound desc = could not find container \"d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2\": container with ID starting with d92b50cdb979d3f7c551355aec315118539f3560c2501c19f5ba1eefc5dc18d2 not found: ID does not exist" Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.083666 4876 scope.go:117] "RemoveContainer" containerID="6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8" Mar 13 08:27:09 crc kubenswrapper[4876]: E0313 08:27:09.084146 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8\": container with ID starting with 6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8 not found: ID does not exist" containerID="6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8" Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.084290 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8"} err="failed to get container status \"6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8\": rpc error: code = NotFound desc = could not find container \"6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8\": container with ID starting with 6299d7f6aa9142f37f22e8fa76a1ca5f1b0efa1416f615ab209cbb9e9bbaf9a8 not found: ID does not exist" Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.084316 4876 scope.go:117] "RemoveContainer" containerID="d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e" Mar 13 08:27:09 crc kubenswrapper[4876]: E0313 08:27:09.084653 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e\": container with ID starting with d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e not found: ID does not exist" containerID="d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e" Mar 13 08:27:09 crc kubenswrapper[4876]: I0313 08:27:09.084716 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e"} err="failed to get container status \"d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e\": rpc error: code = NotFound desc = could not find container \"d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e\": container with ID starting with d2277f47a6885c3232e567f72b61b807a63a6bda1c79f5af84bf7d864188de7e not found: ID does not exist" Mar 13 08:27:11 crc kubenswrapper[4876]: I0313 08:27:11.046433 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" path="/var/lib/kubelet/pods/1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd/volumes" Mar 13 08:27:54 crc kubenswrapper[4876]: I0313 08:27:54.608206 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:27:54 crc kubenswrapper[4876]: I0313 08:27:54.608728 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.201221 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556508-zp5bt"] Mar 13 08:28:00 crc kubenswrapper[4876]: E0313 08:28:00.203078 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerName="extract-utilities" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.203105 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerName="extract-utilities" Mar 13 08:28:00 crc kubenswrapper[4876]: E0313 08:28:00.203136 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerName="extract-content" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.203143 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerName="extract-content" Mar 13 08:28:00 crc kubenswrapper[4876]: E0313 08:28:00.203169 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerName="registry-server" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.203177 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerName="registry-server" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.203529 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b1ad6b6-fd8e-4be4-9b5a-55185fb447dd" containerName="registry-server" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.210139 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556508-zp5bt" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.213864 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556508-zp5bt"] Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.214871 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.215472 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.215954 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.289709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzpgs\" (UniqueName: \"kubernetes.io/projected/86ec1034-717e-4b83-a8c5-5c6b53987942-kube-api-access-bzpgs\") pod \"auto-csr-approver-29556508-zp5bt\" (UID: \"86ec1034-717e-4b83-a8c5-5c6b53987942\") " pod="openshift-infra/auto-csr-approver-29556508-zp5bt" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.393065 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzpgs\" (UniqueName: \"kubernetes.io/projected/86ec1034-717e-4b83-a8c5-5c6b53987942-kube-api-access-bzpgs\") pod \"auto-csr-approver-29556508-zp5bt\" (UID: \"86ec1034-717e-4b83-a8c5-5c6b53987942\") " pod="openshift-infra/auto-csr-approver-29556508-zp5bt" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.426761 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzpgs\" (UniqueName: \"kubernetes.io/projected/86ec1034-717e-4b83-a8c5-5c6b53987942-kube-api-access-bzpgs\") pod \"auto-csr-approver-29556508-zp5bt\" (UID: \"86ec1034-717e-4b83-a8c5-5c6b53987942\") " pod="openshift-infra/auto-csr-approver-29556508-zp5bt" Mar 13 08:28:00 crc kubenswrapper[4876]: I0313 08:28:00.543262 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556508-zp5bt" Mar 13 08:28:01 crc kubenswrapper[4876]: I0313 08:28:01.016158 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556508-zp5bt"] Mar 13 08:28:01 crc kubenswrapper[4876]: I0313 08:28:01.469011 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556508-zp5bt" event={"ID":"86ec1034-717e-4b83-a8c5-5c6b53987942","Type":"ContainerStarted","Data":"dad184ca6e3855d35fbb8eff8a5f2affcd1afb9fc981a247ab89a0db13c271de"} Mar 13 08:28:02 crc kubenswrapper[4876]: I0313 08:28:02.480438 4876 generic.go:334] "Generic (PLEG): container finished" podID="86ec1034-717e-4b83-a8c5-5c6b53987942" containerID="630cf1e046cb38e12ef0b06bddd7d84d8fafddd15e92b03087a60ceeafce927e" exitCode=0 Mar 13 08:28:02 crc kubenswrapper[4876]: I0313 08:28:02.480644 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556508-zp5bt" event={"ID":"86ec1034-717e-4b83-a8c5-5c6b53987942","Type":"ContainerDied","Data":"630cf1e046cb38e12ef0b06bddd7d84d8fafddd15e92b03087a60ceeafce927e"} Mar 13 08:28:04 crc kubenswrapper[4876]: I0313 08:28:03.930964 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556508-zp5bt" Mar 13 08:28:04 crc kubenswrapper[4876]: I0313 08:28:04.080839 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzpgs\" (UniqueName: \"kubernetes.io/projected/86ec1034-717e-4b83-a8c5-5c6b53987942-kube-api-access-bzpgs\") pod \"86ec1034-717e-4b83-a8c5-5c6b53987942\" (UID: \"86ec1034-717e-4b83-a8c5-5c6b53987942\") " Mar 13 08:28:04 crc kubenswrapper[4876]: I0313 08:28:04.098007 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ec1034-717e-4b83-a8c5-5c6b53987942-kube-api-access-bzpgs" (OuterVolumeSpecName: "kube-api-access-bzpgs") pod "86ec1034-717e-4b83-a8c5-5c6b53987942" (UID: "86ec1034-717e-4b83-a8c5-5c6b53987942"). InnerVolumeSpecName "kube-api-access-bzpgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:28:04 crc kubenswrapper[4876]: I0313 08:28:04.183894 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzpgs\" (UniqueName: \"kubernetes.io/projected/86ec1034-717e-4b83-a8c5-5c6b53987942-kube-api-access-bzpgs\") on node \"crc\" DevicePath \"\"" Mar 13 08:28:04 crc kubenswrapper[4876]: I0313 08:28:04.545308 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556508-zp5bt" event={"ID":"86ec1034-717e-4b83-a8c5-5c6b53987942","Type":"ContainerDied","Data":"dad184ca6e3855d35fbb8eff8a5f2affcd1afb9fc981a247ab89a0db13c271de"} Mar 13 08:28:04 crc kubenswrapper[4876]: I0313 08:28:04.545359 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556508-zp5bt" Mar 13 08:28:04 crc kubenswrapper[4876]: I0313 08:28:04.545367 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dad184ca6e3855d35fbb8eff8a5f2affcd1afb9fc981a247ab89a0db13c271de" Mar 13 08:28:05 crc kubenswrapper[4876]: I0313 08:28:05.011592 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556502-pq9vk"] Mar 13 08:28:05 crc kubenswrapper[4876]: I0313 08:28:05.021401 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556502-pq9vk"] Mar 13 08:28:05 crc kubenswrapper[4876]: I0313 08:28:05.047485 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0110e2c-4de2-4bc8-bf9a-fc8745032724" path="/var/lib/kubelet/pods/e0110e2c-4de2-4bc8-bf9a-fc8745032724/volumes" Mar 13 08:28:23 crc kubenswrapper[4876]: I0313 08:28:23.722148 4876 scope.go:117] "RemoveContainer" containerID="ef1f57654297a7a7b06512607d8849db7960efc9e3e058ac36e25ea37154b4b5" Mar 13 08:28:24 crc kubenswrapper[4876]: I0313 08:28:24.607561 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:28:24 crc kubenswrapper[4876]: I0313 08:28:24.607644 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:28:54 crc kubenswrapper[4876]: I0313 08:28:54.607930 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:28:54 crc kubenswrapper[4876]: I0313 08:28:54.608676 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:28:54 crc kubenswrapper[4876]: I0313 08:28:54.608770 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:28:54 crc kubenswrapper[4876]: I0313 08:28:54.609777 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:28:54 crc kubenswrapper[4876]: I0313 08:28:54.609868 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" gracePeriod=600 Mar 13 08:28:54 crc kubenswrapper[4876]: E0313 08:28:54.751407 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:28:55 crc kubenswrapper[4876]: I0313 08:28:55.065357 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" exitCode=0 Mar 13 08:28:55 crc kubenswrapper[4876]: I0313 08:28:55.065558 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b"} Mar 13 08:28:55 crc kubenswrapper[4876]: I0313 08:28:55.066056 4876 scope.go:117] "RemoveContainer" containerID="87c133791187037b4f20020f848722bad2719b25256974c3c1a18083148ed7f2" Mar 13 08:28:55 crc kubenswrapper[4876]: I0313 08:28:55.066982 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:28:55 crc kubenswrapper[4876]: E0313 08:28:55.067368 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:29:07 crc kubenswrapper[4876]: I0313 08:29:07.035442 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:29:07 crc kubenswrapper[4876]: E0313 08:29:07.036361 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:29:20 crc kubenswrapper[4876]: I0313 08:29:20.035967 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:29:20 crc kubenswrapper[4876]: E0313 08:29:20.036919 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:29:35 crc kubenswrapper[4876]: I0313 08:29:35.036102 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:29:35 crc kubenswrapper[4876]: E0313 08:29:35.037094 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:29:50 crc kubenswrapper[4876]: I0313 08:29:50.035642 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:29:50 crc kubenswrapper[4876]: E0313 08:29:50.036923 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.164503 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk"] Mar 13 08:30:00 crc kubenswrapper[4876]: E0313 08:30:00.165649 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ec1034-717e-4b83-a8c5-5c6b53987942" containerName="oc" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.165667 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ec1034-717e-4b83-a8c5-5c6b53987942" containerName="oc" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.165979 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ec1034-717e-4b83-a8c5-5c6b53987942" containerName="oc" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.167055 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.168851 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.171349 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.176885 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk"] Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.267097 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556510-96s8s"] Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.268489 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556510-96s8s" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.273679 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.273952 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.274494 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.291365 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556510-96s8s"] Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.354413 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-config-volume\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.354794 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wj4b\" (UniqueName: \"kubernetes.io/projected/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-kube-api-access-9wj4b\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.354907 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-secret-volume\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.456383 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5f7n\" (UniqueName: \"kubernetes.io/projected/9cab4aa9-1c76-48b3-b8cc-14fc1c07820b-kube-api-access-t5f7n\") pod \"auto-csr-approver-29556510-96s8s\" (UID: \"9cab4aa9-1c76-48b3-b8cc-14fc1c07820b\") " pod="openshift-infra/auto-csr-approver-29556510-96s8s" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.456481 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wj4b\" (UniqueName: \"kubernetes.io/projected/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-kube-api-access-9wj4b\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.456510 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-secret-volume\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.456611 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-config-volume\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.457580 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-config-volume\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.477015 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-secret-volume\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.488606 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wj4b\" (UniqueName: \"kubernetes.io/projected/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-kube-api-access-9wj4b\") pod \"collect-profiles-29556510-pjcwk\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.512879 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.558794 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5f7n\" (UniqueName: \"kubernetes.io/projected/9cab4aa9-1c76-48b3-b8cc-14fc1c07820b-kube-api-access-t5f7n\") pod \"auto-csr-approver-29556510-96s8s\" (UID: \"9cab4aa9-1c76-48b3-b8cc-14fc1c07820b\") " pod="openshift-infra/auto-csr-approver-29556510-96s8s" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.579545 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5f7n\" (UniqueName: \"kubernetes.io/projected/9cab4aa9-1c76-48b3-b8cc-14fc1c07820b-kube-api-access-t5f7n\") pod \"auto-csr-approver-29556510-96s8s\" (UID: \"9cab4aa9-1c76-48b3-b8cc-14fc1c07820b\") " pod="openshift-infra/auto-csr-approver-29556510-96s8s" Mar 13 08:30:00 crc kubenswrapper[4876]: I0313 08:30:00.594016 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556510-96s8s" Mar 13 08:30:01 crc kubenswrapper[4876]: I0313 08:30:01.006524 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk"] Mar 13 08:30:01 crc kubenswrapper[4876]: I0313 08:30:01.133493 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556510-96s8s"] Mar 13 08:30:01 crc kubenswrapper[4876]: W0313 08:30:01.135946 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cab4aa9_1c76_48b3_b8cc_14fc1c07820b.slice/crio-4d0569d549891f6cff97420d19c07a80f017be7126e9a372d6f804d2bb525c29 WatchSource:0}: Error finding container 4d0569d549891f6cff97420d19c07a80f017be7126e9a372d6f804d2bb525c29: Status 404 returned error can't find the container with id 4d0569d549891f6cff97420d19c07a80f017be7126e9a372d6f804d2bb525c29 Mar 13 08:30:01 crc kubenswrapper[4876]: I0313 08:30:01.724756 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" event={"ID":"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd","Type":"ContainerStarted","Data":"b44e6a0e8f7df44d9fc86c1c2491718ecb8ec522cbace37de9b63017276ced49"} Mar 13 08:30:01 crc kubenswrapper[4876]: I0313 08:30:01.725101 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" event={"ID":"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd","Type":"ContainerStarted","Data":"037020f3ff16787df31d3bb62547d0319094e2b75a1be77b0ab270f94b184615"} Mar 13 08:30:01 crc kubenswrapper[4876]: I0313 08:30:01.727436 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556510-96s8s" event={"ID":"9cab4aa9-1c76-48b3-b8cc-14fc1c07820b","Type":"ContainerStarted","Data":"4d0569d549891f6cff97420d19c07a80f017be7126e9a372d6f804d2bb525c29"} Mar 13 08:30:02 crc kubenswrapper[4876]: I0313 08:30:02.036045 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:30:02 crc kubenswrapper[4876]: E0313 08:30:02.036780 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:30:02 crc kubenswrapper[4876]: I0313 08:30:02.769124 4876 generic.go:334] "Generic (PLEG): container finished" podID="0e9877be-ad4f-4ad5-ad48-95e81d30ddfd" containerID="b44e6a0e8f7df44d9fc86c1c2491718ecb8ec522cbace37de9b63017276ced49" exitCode=0 Mar 13 08:30:02 crc kubenswrapper[4876]: I0313 08:30:02.769617 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" event={"ID":"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd","Type":"ContainerDied","Data":"b44e6a0e8f7df44d9fc86c1c2491718ecb8ec522cbace37de9b63017276ced49"} Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.123371 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.316172 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-secret-volume\") pod \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.316277 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-config-volume\") pod \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.316410 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wj4b\" (UniqueName: \"kubernetes.io/projected/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-kube-api-access-9wj4b\") pod \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\" (UID: \"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd\") " Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.317177 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-config-volume" (OuterVolumeSpecName: "config-volume") pod "0e9877be-ad4f-4ad5-ad48-95e81d30ddfd" (UID: "0e9877be-ad4f-4ad5-ad48-95e81d30ddfd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.322748 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0e9877be-ad4f-4ad5-ad48-95e81d30ddfd" (UID: "0e9877be-ad4f-4ad5-ad48-95e81d30ddfd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.323109 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-kube-api-access-9wj4b" (OuterVolumeSpecName: "kube-api-access-9wj4b") pod "0e9877be-ad4f-4ad5-ad48-95e81d30ddfd" (UID: "0e9877be-ad4f-4ad5-ad48-95e81d30ddfd"). InnerVolumeSpecName "kube-api-access-9wj4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.418473 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wj4b\" (UniqueName: \"kubernetes.io/projected/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-kube-api-access-9wj4b\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.418872 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.418888 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e9877be-ad4f-4ad5-ad48-95e81d30ddfd-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.778576 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" event={"ID":"0e9877be-ad4f-4ad5-ad48-95e81d30ddfd","Type":"ContainerDied","Data":"037020f3ff16787df31d3bb62547d0319094e2b75a1be77b0ab270f94b184615"} Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.778644 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="037020f3ff16787df31d3bb62547d0319094e2b75a1be77b0ab270f94b184615" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.778646 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556510-pjcwk" Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.780569 4876 generic.go:334] "Generic (PLEG): container finished" podID="9cab4aa9-1c76-48b3-b8cc-14fc1c07820b" containerID="a77192805bafccf47da7f79e035324dc4fdd928380259b1d34da6a9f1f339732" exitCode=0 Mar 13 08:30:03 crc kubenswrapper[4876]: I0313 08:30:03.780597 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556510-96s8s" event={"ID":"9cab4aa9-1c76-48b3-b8cc-14fc1c07820b","Type":"ContainerDied","Data":"a77192805bafccf47da7f79e035324dc4fdd928380259b1d34da6a9f1f339732"} Mar 13 08:30:04 crc kubenswrapper[4876]: I0313 08:30:04.207359 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6"] Mar 13 08:30:04 crc kubenswrapper[4876]: I0313 08:30:04.214688 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556465-pjcv6"] Mar 13 08:30:05 crc kubenswrapper[4876]: I0313 08:30:05.054950 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="884d6b32-5128-4103-a9b6-82fb23d792ce" path="/var/lib/kubelet/pods/884d6b32-5128-4103-a9b6-82fb23d792ce/volumes" Mar 13 08:30:05 crc kubenswrapper[4876]: I0313 08:30:05.353437 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556510-96s8s" Mar 13 08:30:05 crc kubenswrapper[4876]: I0313 08:30:05.397077 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5f7n\" (UniqueName: \"kubernetes.io/projected/9cab4aa9-1c76-48b3-b8cc-14fc1c07820b-kube-api-access-t5f7n\") pod \"9cab4aa9-1c76-48b3-b8cc-14fc1c07820b\" (UID: \"9cab4aa9-1c76-48b3-b8cc-14fc1c07820b\") " Mar 13 08:30:05 crc kubenswrapper[4876]: I0313 08:30:05.409212 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cab4aa9-1c76-48b3-b8cc-14fc1c07820b-kube-api-access-t5f7n" (OuterVolumeSpecName: "kube-api-access-t5f7n") pod "9cab4aa9-1c76-48b3-b8cc-14fc1c07820b" (UID: "9cab4aa9-1c76-48b3-b8cc-14fc1c07820b"). InnerVolumeSpecName "kube-api-access-t5f7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:30:05 crc kubenswrapper[4876]: I0313 08:30:05.499083 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5f7n\" (UniqueName: \"kubernetes.io/projected/9cab4aa9-1c76-48b3-b8cc-14fc1c07820b-kube-api-access-t5f7n\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:05 crc kubenswrapper[4876]: I0313 08:30:05.795667 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556510-96s8s" event={"ID":"9cab4aa9-1c76-48b3-b8cc-14fc1c07820b","Type":"ContainerDied","Data":"4d0569d549891f6cff97420d19c07a80f017be7126e9a372d6f804d2bb525c29"} Mar 13 08:30:05 crc kubenswrapper[4876]: I0313 08:30:05.795704 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d0569d549891f6cff97420d19c07a80f017be7126e9a372d6f804d2bb525c29" Mar 13 08:30:05 crc kubenswrapper[4876]: I0313 08:30:05.795753 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556510-96s8s" Mar 13 08:30:06 crc kubenswrapper[4876]: I0313 08:30:06.410884 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556504-2mrbk"] Mar 13 08:30:06 crc kubenswrapper[4876]: I0313 08:30:06.428250 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556504-2mrbk"] Mar 13 08:30:07 crc kubenswrapper[4876]: I0313 08:30:07.047115 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eba926b-7bb7-4522-bc87-94f5ea0ad6f4" path="/var/lib/kubelet/pods/0eba926b-7bb7-4522-bc87-94f5ea0ad6f4/volumes" Mar 13 08:30:15 crc kubenswrapper[4876]: I0313 08:30:15.036768 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:30:15 crc kubenswrapper[4876]: E0313 08:30:15.037839 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:30:23 crc kubenswrapper[4876]: I0313 08:30:23.834858 4876 scope.go:117] "RemoveContainer" containerID="8cbcd17c6bff74394c479fba13f47c4e28fb41c91324bde5455471f54d1a7053" Mar 13 08:30:23 crc kubenswrapper[4876]: I0313 08:30:23.907232 4876 scope.go:117] "RemoveContainer" containerID="edef007bd0bcea1e3c47cd9810ded713471e6f3dd57dbd7814a0aa908b17dc62" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.036570 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:30:27 crc kubenswrapper[4876]: E0313 08:30:27.037779 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.506916 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pkbcs"] Mar 13 08:30:27 crc kubenswrapper[4876]: E0313 08:30:27.507304 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e9877be-ad4f-4ad5-ad48-95e81d30ddfd" containerName="collect-profiles" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.507318 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e9877be-ad4f-4ad5-ad48-95e81d30ddfd" containerName="collect-profiles" Mar 13 08:30:27 crc kubenswrapper[4876]: E0313 08:30:27.507333 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cab4aa9-1c76-48b3-b8cc-14fc1c07820b" containerName="oc" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.507339 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cab4aa9-1c76-48b3-b8cc-14fc1c07820b" containerName="oc" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.507518 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e9877be-ad4f-4ad5-ad48-95e81d30ddfd" containerName="collect-profiles" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.507534 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cab4aa9-1c76-48b3-b8cc-14fc1c07820b" containerName="oc" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.508995 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.541744 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pkbcs"] Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.676971 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-285pm\" (UniqueName: \"kubernetes.io/projected/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-kube-api-access-285pm\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.677047 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-utilities\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.677216 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-catalog-content\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.779940 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-utilities\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.780021 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-catalog-content\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.780149 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-285pm\" (UniqueName: \"kubernetes.io/projected/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-kube-api-access-285pm\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.780683 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-catalog-content\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.781055 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-utilities\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.809885 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-285pm\" (UniqueName: \"kubernetes.io/projected/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-kube-api-access-285pm\") pod \"community-operators-pkbcs\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:27 crc kubenswrapper[4876]: I0313 08:30:27.852158 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:28 crc kubenswrapper[4876]: I0313 08:30:28.393033 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pkbcs"] Mar 13 08:30:29 crc kubenswrapper[4876]: I0313 08:30:29.070804 4876 generic.go:334] "Generic (PLEG): container finished" podID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerID="90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147" exitCode=0 Mar 13 08:30:29 crc kubenswrapper[4876]: I0313 08:30:29.070865 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkbcs" event={"ID":"175cf39c-7cf6-4868-bea2-3dc2680a9ee3","Type":"ContainerDied","Data":"90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147"} Mar 13 08:30:29 crc kubenswrapper[4876]: I0313 08:30:29.071323 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkbcs" event={"ID":"175cf39c-7cf6-4868-bea2-3dc2680a9ee3","Type":"ContainerStarted","Data":"48d45d53b619c5b3100dd53db8e1a68823c7601ca3833820293337e74da4f801"} Mar 13 08:30:29 crc kubenswrapper[4876]: I0313 08:30:29.903932 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mbhqb"] Mar 13 08:30:29 crc kubenswrapper[4876]: I0313 08:30:29.907036 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:29 crc kubenswrapper[4876]: I0313 08:30:29.923953 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbhqb"] Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.058142 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2k8q\" (UniqueName: \"kubernetes.io/projected/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-kube-api-access-n2k8q\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.058259 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-catalog-content\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.058308 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-utilities\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.083311 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkbcs" event={"ID":"175cf39c-7cf6-4868-bea2-3dc2680a9ee3","Type":"ContainerStarted","Data":"eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52"} Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.160904 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-catalog-content\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.160958 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-utilities\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.161150 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2k8q\" (UniqueName: \"kubernetes.io/projected/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-kube-api-access-n2k8q\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.161993 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-catalog-content\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.162328 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-utilities\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.182837 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2k8q\" (UniqueName: \"kubernetes.io/projected/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-kube-api-access-n2k8q\") pod \"certified-operators-mbhqb\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.243189 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:30 crc kubenswrapper[4876]: I0313 08:30:30.757917 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbhqb"] Mar 13 08:30:30 crc kubenswrapper[4876]: W0313 08:30:30.767548 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb85b45e9_4401_4ebf_b6f1_8f0f6d947220.slice/crio-cb1a34babd8261c15746f659e698c3078b0a25140f108446a110911c0d34a277 WatchSource:0}: Error finding container cb1a34babd8261c15746f659e698c3078b0a25140f108446a110911c0d34a277: Status 404 returned error can't find the container with id cb1a34babd8261c15746f659e698c3078b0a25140f108446a110911c0d34a277 Mar 13 08:30:31 crc kubenswrapper[4876]: I0313 08:30:31.095255 4876 generic.go:334] "Generic (PLEG): container finished" podID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerID="69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1" exitCode=0 Mar 13 08:30:31 crc kubenswrapper[4876]: I0313 08:30:31.095971 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbhqb" event={"ID":"b85b45e9-4401-4ebf-b6f1-8f0f6d947220","Type":"ContainerDied","Data":"69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1"} Mar 13 08:30:31 crc kubenswrapper[4876]: I0313 08:30:31.096031 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbhqb" event={"ID":"b85b45e9-4401-4ebf-b6f1-8f0f6d947220","Type":"ContainerStarted","Data":"cb1a34babd8261c15746f659e698c3078b0a25140f108446a110911c0d34a277"} Mar 13 08:30:31 crc kubenswrapper[4876]: I0313 08:30:31.098803 4876 generic.go:334] "Generic (PLEG): container finished" podID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerID="eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52" exitCode=0 Mar 13 08:30:31 crc kubenswrapper[4876]: I0313 08:30:31.098858 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkbcs" event={"ID":"175cf39c-7cf6-4868-bea2-3dc2680a9ee3","Type":"ContainerDied","Data":"eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52"} Mar 13 08:30:32 crc kubenswrapper[4876]: I0313 08:30:32.112627 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkbcs" event={"ID":"175cf39c-7cf6-4868-bea2-3dc2680a9ee3","Type":"ContainerStarted","Data":"37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb"} Mar 13 08:30:32 crc kubenswrapper[4876]: I0313 08:30:32.586659 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pkbcs" podStartSLOduration=3.092061196 podStartE2EDuration="5.586624264s" podCreationTimestamp="2026-03-13 08:30:27 +0000 UTC" firstStartedPulling="2026-03-13 08:30:29.073695444 +0000 UTC m=+3088.744474426" lastFinishedPulling="2026-03-13 08:30:31.568258512 +0000 UTC m=+3091.239037494" observedRunningTime="2026-03-13 08:30:32.581341151 +0000 UTC m=+3092.252120133" watchObservedRunningTime="2026-03-13 08:30:32.586624264 +0000 UTC m=+3092.257403286" Mar 13 08:30:33 crc kubenswrapper[4876]: I0313 08:30:33.127602 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbhqb" event={"ID":"b85b45e9-4401-4ebf-b6f1-8f0f6d947220","Type":"ContainerStarted","Data":"2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1"} Mar 13 08:30:34 crc kubenswrapper[4876]: I0313 08:30:34.143035 4876 generic.go:334] "Generic (PLEG): container finished" podID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerID="2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1" exitCode=0 Mar 13 08:30:34 crc kubenswrapper[4876]: I0313 08:30:34.143104 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbhqb" event={"ID":"b85b45e9-4401-4ebf-b6f1-8f0f6d947220","Type":"ContainerDied","Data":"2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1"} Mar 13 08:30:35 crc kubenswrapper[4876]: I0313 08:30:35.166978 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbhqb" event={"ID":"b85b45e9-4401-4ebf-b6f1-8f0f6d947220","Type":"ContainerStarted","Data":"c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d"} Mar 13 08:30:37 crc kubenswrapper[4876]: I0313 08:30:37.853409 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:37 crc kubenswrapper[4876]: I0313 08:30:37.854291 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:37 crc kubenswrapper[4876]: I0313 08:30:37.935772 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:37 crc kubenswrapper[4876]: I0313 08:30:37.977607 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mbhqb" podStartSLOduration=5.195629098 podStartE2EDuration="8.977557456s" podCreationTimestamp="2026-03-13 08:30:29 +0000 UTC" firstStartedPulling="2026-03-13 08:30:31.097504048 +0000 UTC m=+3090.768283040" lastFinishedPulling="2026-03-13 08:30:34.879432406 +0000 UTC m=+3094.550211398" observedRunningTime="2026-03-13 08:30:35.188296389 +0000 UTC m=+3094.859075371" watchObservedRunningTime="2026-03-13 08:30:37.977557456 +0000 UTC m=+3097.648336468" Mar 13 08:30:38 crc kubenswrapper[4876]: I0313 08:30:38.036947 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:30:38 crc kubenswrapper[4876]: E0313 08:30:38.037296 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:30:38 crc kubenswrapper[4876]: I0313 08:30:38.252032 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.090931 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pkbcs"] Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.214088 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pkbcs" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerName="registry-server" containerID="cri-o://37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb" gracePeriod=2 Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.243298 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.245549 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.814206 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.847732 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-utilities\") pod \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.848178 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-catalog-content\") pod \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.848247 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-285pm\" (UniqueName: \"kubernetes.io/projected/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-kube-api-access-285pm\") pod \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\" (UID: \"175cf39c-7cf6-4868-bea2-3dc2680a9ee3\") " Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.848396 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-utilities" (OuterVolumeSpecName: "utilities") pod "175cf39c-7cf6-4868-bea2-3dc2680a9ee3" (UID: "175cf39c-7cf6-4868-bea2-3dc2680a9ee3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.848832 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.857608 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-kube-api-access-285pm" (OuterVolumeSpecName: "kube-api-access-285pm") pod "175cf39c-7cf6-4868-bea2-3dc2680a9ee3" (UID: "175cf39c-7cf6-4868-bea2-3dc2680a9ee3"). InnerVolumeSpecName "kube-api-access-285pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.903014 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "175cf39c-7cf6-4868-bea2-3dc2680a9ee3" (UID: "175cf39c-7cf6-4868-bea2-3dc2680a9ee3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.952211 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:40 crc kubenswrapper[4876]: I0313 08:30:40.952292 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-285pm\" (UniqueName: \"kubernetes.io/projected/175cf39c-7cf6-4868-bea2-3dc2680a9ee3-kube-api-access-285pm\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.230301 4876 generic.go:334] "Generic (PLEG): container finished" podID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerID="37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb" exitCode=0 Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.230389 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pkbcs" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.230443 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkbcs" event={"ID":"175cf39c-7cf6-4868-bea2-3dc2680a9ee3","Type":"ContainerDied","Data":"37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb"} Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.230475 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkbcs" event={"ID":"175cf39c-7cf6-4868-bea2-3dc2680a9ee3","Type":"ContainerDied","Data":"48d45d53b619c5b3100dd53db8e1a68823c7601ca3833820293337e74da4f801"} Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.230498 4876 scope.go:117] "RemoveContainer" containerID="37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.268911 4876 scope.go:117] "RemoveContainer" containerID="eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.271672 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pkbcs"] Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.294841 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pkbcs"] Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.300285 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mbhqb" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="registry-server" probeResult="failure" output=< Mar 13 08:30:41 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:30:41 crc kubenswrapper[4876]: > Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.301995 4876 scope.go:117] "RemoveContainer" containerID="90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.340318 4876 scope.go:117] "RemoveContainer" containerID="37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb" Mar 13 08:30:41 crc kubenswrapper[4876]: E0313 08:30:41.340776 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb\": container with ID starting with 37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb not found: ID does not exist" containerID="37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.340806 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb"} err="failed to get container status \"37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb\": rpc error: code = NotFound desc = could not find container \"37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb\": container with ID starting with 37153f3beb68af67b956bd508bbfbeeae8acde8785bd7293331c34eddb02abdb not found: ID does not exist" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.340825 4876 scope.go:117] "RemoveContainer" containerID="eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52" Mar 13 08:30:41 crc kubenswrapper[4876]: E0313 08:30:41.341096 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52\": container with ID starting with eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52 not found: ID does not exist" containerID="eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.341115 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52"} err="failed to get container status \"eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52\": rpc error: code = NotFound desc = could not find container \"eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52\": container with ID starting with eae73818f2f914c2c19baae684830dbec6180bd9bf4d110a9f2494d9f6b76a52 not found: ID does not exist" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.341129 4876 scope.go:117] "RemoveContainer" containerID="90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147" Mar 13 08:30:41 crc kubenswrapper[4876]: E0313 08:30:41.341375 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147\": container with ID starting with 90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147 not found: ID does not exist" containerID="90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147" Mar 13 08:30:41 crc kubenswrapper[4876]: I0313 08:30:41.341394 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147"} err="failed to get container status \"90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147\": rpc error: code = NotFound desc = could not find container \"90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147\": container with ID starting with 90007f19d9c8bbfcde5c1b5ead9e60b8d1ab0340dd8ceebb549883c61e574147 not found: ID does not exist" Mar 13 08:30:43 crc kubenswrapper[4876]: I0313 08:30:43.049632 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" path="/var/lib/kubelet/pods/175cf39c-7cf6-4868-bea2-3dc2680a9ee3/volumes" Mar 13 08:30:50 crc kubenswrapper[4876]: I0313 08:30:50.302227 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:50 crc kubenswrapper[4876]: I0313 08:30:50.367154 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:50 crc kubenswrapper[4876]: I0313 08:30:50.561133 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbhqb"] Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.045061 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:30:51 crc kubenswrapper[4876]: E0313 08:30:51.045831 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.347606 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mbhqb" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="registry-server" containerID="cri-o://c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d" gracePeriod=2 Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.852737 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.923552 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2k8q\" (UniqueName: \"kubernetes.io/projected/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-kube-api-access-n2k8q\") pod \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.925035 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-utilities\") pod \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.925263 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-catalog-content\") pod \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\" (UID: \"b85b45e9-4401-4ebf-b6f1-8f0f6d947220\") " Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.926141 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-utilities" (OuterVolumeSpecName: "utilities") pod "b85b45e9-4401-4ebf-b6f1-8f0f6d947220" (UID: "b85b45e9-4401-4ebf-b6f1-8f0f6d947220"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.926460 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.938074 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-kube-api-access-n2k8q" (OuterVolumeSpecName: "kube-api-access-n2k8q") pod "b85b45e9-4401-4ebf-b6f1-8f0f6d947220" (UID: "b85b45e9-4401-4ebf-b6f1-8f0f6d947220"). InnerVolumeSpecName "kube-api-access-n2k8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:30:51 crc kubenswrapper[4876]: I0313 08:30:51.992142 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b85b45e9-4401-4ebf-b6f1-8f0f6d947220" (UID: "b85b45e9-4401-4ebf-b6f1-8f0f6d947220"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.028097 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2k8q\" (UniqueName: \"kubernetes.io/projected/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-kube-api-access-n2k8q\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.028136 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85b45e9-4401-4ebf-b6f1-8f0f6d947220-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.363697 4876 generic.go:334] "Generic (PLEG): container finished" podID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerID="c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d" exitCode=0 Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.363746 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbhqb" event={"ID":"b85b45e9-4401-4ebf-b6f1-8f0f6d947220","Type":"ContainerDied","Data":"c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d"} Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.363773 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbhqb" event={"ID":"b85b45e9-4401-4ebf-b6f1-8f0f6d947220","Type":"ContainerDied","Data":"cb1a34babd8261c15746f659e698c3078b0a25140f108446a110911c0d34a277"} Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.363792 4876 scope.go:117] "RemoveContainer" containerID="c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.363920 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbhqb" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.403885 4876 scope.go:117] "RemoveContainer" containerID="2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.405269 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbhqb"] Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.413771 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mbhqb"] Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.452764 4876 scope.go:117] "RemoveContainer" containerID="69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.489453 4876 scope.go:117] "RemoveContainer" containerID="c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d" Mar 13 08:30:52 crc kubenswrapper[4876]: E0313 08:30:52.490123 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d\": container with ID starting with c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d not found: ID does not exist" containerID="c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.490158 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d"} err="failed to get container status \"c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d\": rpc error: code = NotFound desc = could not find container \"c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d\": container with ID starting with c641e2fcdf3ce04517192840c862c72f720118d5217ae9717cc82a80bf2f4f5d not found: ID does not exist" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.490181 4876 scope.go:117] "RemoveContainer" containerID="2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1" Mar 13 08:30:52 crc kubenswrapper[4876]: E0313 08:30:52.490672 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1\": container with ID starting with 2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1 not found: ID does not exist" containerID="2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.490744 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1"} err="failed to get container status \"2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1\": rpc error: code = NotFound desc = could not find container \"2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1\": container with ID starting with 2bb2ad49570066dbbebb0a8e6466c00a921c05f3ae6964f0a6a423625189f0a1 not found: ID does not exist" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.490776 4876 scope.go:117] "RemoveContainer" containerID="69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1" Mar 13 08:30:52 crc kubenswrapper[4876]: E0313 08:30:52.491126 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1\": container with ID starting with 69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1 not found: ID does not exist" containerID="69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1" Mar 13 08:30:52 crc kubenswrapper[4876]: I0313 08:30:52.491169 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1"} err="failed to get container status \"69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1\": rpc error: code = NotFound desc = could not find container \"69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1\": container with ID starting with 69f18e163a6e4c4531c22d2b994ab80d28f3c091805de43648606ee3cc528cc1 not found: ID does not exist" Mar 13 08:30:53 crc kubenswrapper[4876]: I0313 08:30:53.049410 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" path="/var/lib/kubelet/pods/b85b45e9-4401-4ebf-b6f1-8f0f6d947220/volumes" Mar 13 08:31:02 crc kubenswrapper[4876]: I0313 08:31:02.035906 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:31:02 crc kubenswrapper[4876]: E0313 08:31:02.036909 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:31:16 crc kubenswrapper[4876]: I0313 08:31:16.036325 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:31:16 crc kubenswrapper[4876]: E0313 08:31:16.037403 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:31:28 crc kubenswrapper[4876]: I0313 08:31:28.035689 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:31:28 crc kubenswrapper[4876]: E0313 08:31:28.036926 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:31:41 crc kubenswrapper[4876]: I0313 08:31:41.041562 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:31:41 crc kubenswrapper[4876]: E0313 08:31:41.042815 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:31:56 crc kubenswrapper[4876]: I0313 08:31:56.036705 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:31:56 crc kubenswrapper[4876]: E0313 08:31:56.037807 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.157856 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556512-z7ps6"] Mar 13 08:32:00 crc kubenswrapper[4876]: E0313 08:32:00.158727 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="registry-server" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.158740 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="registry-server" Mar 13 08:32:00 crc kubenswrapper[4876]: E0313 08:32:00.158751 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerName="extract-utilities" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.158757 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerName="extract-utilities" Mar 13 08:32:00 crc kubenswrapper[4876]: E0313 08:32:00.158776 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="extract-utilities" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.158782 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="extract-utilities" Mar 13 08:32:00 crc kubenswrapper[4876]: E0313 08:32:00.158805 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="extract-content" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.158818 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="extract-content" Mar 13 08:32:00 crc kubenswrapper[4876]: E0313 08:32:00.158844 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerName="extract-content" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.158855 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerName="extract-content" Mar 13 08:32:00 crc kubenswrapper[4876]: E0313 08:32:00.158865 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerName="registry-server" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.158873 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerName="registry-server" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.159090 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="175cf39c-7cf6-4868-bea2-3dc2680a9ee3" containerName="registry-server" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.159110 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85b45e9-4401-4ebf-b6f1-8f0f6d947220" containerName="registry-server" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.159945 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.164188 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.164684 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.165704 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.183471 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556512-z7ps6"] Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.285012 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw42x\" (UniqueName: \"kubernetes.io/projected/7e09b27f-c606-4be9-83c9-152efce10f10-kube-api-access-fw42x\") pod \"auto-csr-approver-29556512-z7ps6\" (UID: \"7e09b27f-c606-4be9-83c9-152efce10f10\") " pod="openshift-infra/auto-csr-approver-29556512-z7ps6" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.386808 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw42x\" (UniqueName: \"kubernetes.io/projected/7e09b27f-c606-4be9-83c9-152efce10f10-kube-api-access-fw42x\") pod \"auto-csr-approver-29556512-z7ps6\" (UID: \"7e09b27f-c606-4be9-83c9-152efce10f10\") " pod="openshift-infra/auto-csr-approver-29556512-z7ps6" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.411058 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw42x\" (UniqueName: \"kubernetes.io/projected/7e09b27f-c606-4be9-83c9-152efce10f10-kube-api-access-fw42x\") pod \"auto-csr-approver-29556512-z7ps6\" (UID: \"7e09b27f-c606-4be9-83c9-152efce10f10\") " pod="openshift-infra/auto-csr-approver-29556512-z7ps6" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.485491 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.929423 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556512-z7ps6"] Mar 13 08:32:00 crc kubenswrapper[4876]: I0313 08:32:00.940280 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:32:01 crc kubenswrapper[4876]: I0313 08:32:01.160853 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" event={"ID":"7e09b27f-c606-4be9-83c9-152efce10f10","Type":"ContainerStarted","Data":"6bc6b0736d5c40b17d7f96d99b711400881d0470d0fe93d654f30f80462b312b"} Mar 13 08:32:02 crc kubenswrapper[4876]: I0313 08:32:02.170686 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" event={"ID":"7e09b27f-c606-4be9-83c9-152efce10f10","Type":"ContainerStarted","Data":"0cefe81b31844959742d7080e6ac8ccfc509616720396f9ce49c94513fad867a"} Mar 13 08:32:02 crc kubenswrapper[4876]: I0313 08:32:02.197708 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" podStartSLOduration=1.282420533 podStartE2EDuration="2.197689377s" podCreationTimestamp="2026-03-13 08:32:00 +0000 UTC" firstStartedPulling="2026-03-13 08:32:00.940059866 +0000 UTC m=+3180.610838838" lastFinishedPulling="2026-03-13 08:32:01.8553287 +0000 UTC m=+3181.526107682" observedRunningTime="2026-03-13 08:32:02.191264882 +0000 UTC m=+3181.862043864" watchObservedRunningTime="2026-03-13 08:32:02.197689377 +0000 UTC m=+3181.868468359" Mar 13 08:32:03 crc kubenswrapper[4876]: I0313 08:32:03.180291 4876 generic.go:334] "Generic (PLEG): container finished" podID="7e09b27f-c606-4be9-83c9-152efce10f10" containerID="0cefe81b31844959742d7080e6ac8ccfc509616720396f9ce49c94513fad867a" exitCode=0 Mar 13 08:32:03 crc kubenswrapper[4876]: I0313 08:32:03.180345 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" event={"ID":"7e09b27f-c606-4be9-83c9-152efce10f10","Type":"ContainerDied","Data":"0cefe81b31844959742d7080e6ac8ccfc509616720396f9ce49c94513fad867a"} Mar 13 08:32:04 crc kubenswrapper[4876]: I0313 08:32:04.663719 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" Mar 13 08:32:04 crc kubenswrapper[4876]: I0313 08:32:04.767029 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw42x\" (UniqueName: \"kubernetes.io/projected/7e09b27f-c606-4be9-83c9-152efce10f10-kube-api-access-fw42x\") pod \"7e09b27f-c606-4be9-83c9-152efce10f10\" (UID: \"7e09b27f-c606-4be9-83c9-152efce10f10\") " Mar 13 08:32:04 crc kubenswrapper[4876]: I0313 08:32:04.773471 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e09b27f-c606-4be9-83c9-152efce10f10-kube-api-access-fw42x" (OuterVolumeSpecName: "kube-api-access-fw42x") pod "7e09b27f-c606-4be9-83c9-152efce10f10" (UID: "7e09b27f-c606-4be9-83c9-152efce10f10"). InnerVolumeSpecName "kube-api-access-fw42x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:32:04 crc kubenswrapper[4876]: I0313 08:32:04.869180 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw42x\" (UniqueName: \"kubernetes.io/projected/7e09b27f-c606-4be9-83c9-152efce10f10-kube-api-access-fw42x\") on node \"crc\" DevicePath \"\"" Mar 13 08:32:05 crc kubenswrapper[4876]: I0313 08:32:05.208594 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" event={"ID":"7e09b27f-c606-4be9-83c9-152efce10f10","Type":"ContainerDied","Data":"6bc6b0736d5c40b17d7f96d99b711400881d0470d0fe93d654f30f80462b312b"} Mar 13 08:32:05 crc kubenswrapper[4876]: I0313 08:32:05.208637 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bc6b0736d5c40b17d7f96d99b711400881d0470d0fe93d654f30f80462b312b" Mar 13 08:32:05 crc kubenswrapper[4876]: I0313 08:32:05.208652 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556512-z7ps6" Mar 13 08:32:05 crc kubenswrapper[4876]: I0313 08:32:05.258801 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556506-z26rs"] Mar 13 08:32:05 crc kubenswrapper[4876]: I0313 08:32:05.267559 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556506-z26rs"] Mar 13 08:32:07 crc kubenswrapper[4876]: I0313 08:32:07.045940 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7c968c-dac3-4d92-bd12-7334775430c7" path="/var/lib/kubelet/pods/ff7c968c-dac3-4d92-bd12-7334775430c7/volumes" Mar 13 08:32:08 crc kubenswrapper[4876]: I0313 08:32:08.037863 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:32:08 crc kubenswrapper[4876]: E0313 08:32:08.038710 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:32:20 crc kubenswrapper[4876]: I0313 08:32:20.035504 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:32:20 crc kubenswrapper[4876]: E0313 08:32:20.036226 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:32:24 crc kubenswrapper[4876]: I0313 08:32:24.064019 4876 scope.go:117] "RemoveContainer" containerID="f359a0e13dcb77d3e97aa0d981d74018b291a3aaf838877976fca83c5fc98f94" Mar 13 08:32:34 crc kubenswrapper[4876]: I0313 08:32:34.035618 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:32:34 crc kubenswrapper[4876]: E0313 08:32:34.037016 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:32:45 crc kubenswrapper[4876]: I0313 08:32:45.035180 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:32:45 crc kubenswrapper[4876]: E0313 08:32:45.035770 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:33:00 crc kubenswrapper[4876]: I0313 08:33:00.035735 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:33:00 crc kubenswrapper[4876]: E0313 08:33:00.036860 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:33:14 crc kubenswrapper[4876]: I0313 08:33:14.035808 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:33:14 crc kubenswrapper[4876]: E0313 08:33:14.036653 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:33:27 crc kubenswrapper[4876]: I0313 08:33:27.035880 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:33:27 crc kubenswrapper[4876]: E0313 08:33:27.036743 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:33:41 crc kubenswrapper[4876]: I0313 08:33:41.040837 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:33:41 crc kubenswrapper[4876]: E0313 08:33:41.041567 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:33:52 crc kubenswrapper[4876]: I0313 08:33:52.037684 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:33:52 crc kubenswrapper[4876]: E0313 08:33:52.039180 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.217755 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556514-7mcf6"] Mar 13 08:34:00 crc kubenswrapper[4876]: E0313 08:34:00.218653 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e09b27f-c606-4be9-83c9-152efce10f10" containerName="oc" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.218668 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e09b27f-c606-4be9-83c9-152efce10f10" containerName="oc" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.218848 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e09b27f-c606-4be9-83c9-152efce10f10" containerName="oc" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.219466 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556514-7mcf6" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.223404 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.223822 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.226623 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.228974 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmjn6\" (UniqueName: \"kubernetes.io/projected/2d2cee80-a2b8-425d-b58e-54c2710e73a3-kube-api-access-kmjn6\") pod \"auto-csr-approver-29556514-7mcf6\" (UID: \"2d2cee80-a2b8-425d-b58e-54c2710e73a3\") " pod="openshift-infra/auto-csr-approver-29556514-7mcf6" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.239427 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556514-7mcf6"] Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.331292 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmjn6\" (UniqueName: \"kubernetes.io/projected/2d2cee80-a2b8-425d-b58e-54c2710e73a3-kube-api-access-kmjn6\") pod \"auto-csr-approver-29556514-7mcf6\" (UID: \"2d2cee80-a2b8-425d-b58e-54c2710e73a3\") " pod="openshift-infra/auto-csr-approver-29556514-7mcf6" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.352945 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmjn6\" (UniqueName: \"kubernetes.io/projected/2d2cee80-a2b8-425d-b58e-54c2710e73a3-kube-api-access-kmjn6\") pod \"auto-csr-approver-29556514-7mcf6\" (UID: \"2d2cee80-a2b8-425d-b58e-54c2710e73a3\") " pod="openshift-infra/auto-csr-approver-29556514-7mcf6" Mar 13 08:34:00 crc kubenswrapper[4876]: I0313 08:34:00.535788 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556514-7mcf6" Mar 13 08:34:01 crc kubenswrapper[4876]: I0313 08:34:01.014087 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556514-7mcf6"] Mar 13 08:34:01 crc kubenswrapper[4876]: I0313 08:34:01.625020 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556514-7mcf6" event={"ID":"2d2cee80-a2b8-425d-b58e-54c2710e73a3","Type":"ContainerStarted","Data":"6fbd057cbd92974c6943256eff15088eb68f8e8f44845b4ef15f2c4021ab96d7"} Mar 13 08:34:02 crc kubenswrapper[4876]: I0313 08:34:02.644822 4876 generic.go:334] "Generic (PLEG): container finished" podID="2d2cee80-a2b8-425d-b58e-54c2710e73a3" containerID="c51464a0d947ef4d66542e316dd62108881d81eb6294e064097a6bcf21099d02" exitCode=0 Mar 13 08:34:02 crc kubenswrapper[4876]: I0313 08:34:02.644907 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556514-7mcf6" event={"ID":"2d2cee80-a2b8-425d-b58e-54c2710e73a3","Type":"ContainerDied","Data":"c51464a0d947ef4d66542e316dd62108881d81eb6294e064097a6bcf21099d02"} Mar 13 08:34:04 crc kubenswrapper[4876]: I0313 08:34:04.083752 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556514-7mcf6" Mar 13 08:34:04 crc kubenswrapper[4876]: I0313 08:34:04.212000 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmjn6\" (UniqueName: \"kubernetes.io/projected/2d2cee80-a2b8-425d-b58e-54c2710e73a3-kube-api-access-kmjn6\") pod \"2d2cee80-a2b8-425d-b58e-54c2710e73a3\" (UID: \"2d2cee80-a2b8-425d-b58e-54c2710e73a3\") " Mar 13 08:34:04 crc kubenswrapper[4876]: I0313 08:34:04.219356 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d2cee80-a2b8-425d-b58e-54c2710e73a3-kube-api-access-kmjn6" (OuterVolumeSpecName: "kube-api-access-kmjn6") pod "2d2cee80-a2b8-425d-b58e-54c2710e73a3" (UID: "2d2cee80-a2b8-425d-b58e-54c2710e73a3"). InnerVolumeSpecName "kube-api-access-kmjn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:34:04 crc kubenswrapper[4876]: I0313 08:34:04.314476 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmjn6\" (UniqueName: \"kubernetes.io/projected/2d2cee80-a2b8-425d-b58e-54c2710e73a3-kube-api-access-kmjn6\") on node \"crc\" DevicePath \"\"" Mar 13 08:34:04 crc kubenswrapper[4876]: I0313 08:34:04.665345 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556514-7mcf6" event={"ID":"2d2cee80-a2b8-425d-b58e-54c2710e73a3","Type":"ContainerDied","Data":"6fbd057cbd92974c6943256eff15088eb68f8e8f44845b4ef15f2c4021ab96d7"} Mar 13 08:34:04 crc kubenswrapper[4876]: I0313 08:34:04.665390 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556514-7mcf6" Mar 13 08:34:04 crc kubenswrapper[4876]: I0313 08:34:04.665392 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fbd057cbd92974c6943256eff15088eb68f8e8f44845b4ef15f2c4021ab96d7" Mar 13 08:34:05 crc kubenswrapper[4876]: I0313 08:34:05.159168 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556508-zp5bt"] Mar 13 08:34:05 crc kubenswrapper[4876]: I0313 08:34:05.167129 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556508-zp5bt"] Mar 13 08:34:07 crc kubenswrapper[4876]: I0313 08:34:07.036472 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:34:07 crc kubenswrapper[4876]: I0313 08:34:07.049973 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86ec1034-717e-4b83-a8c5-5c6b53987942" path="/var/lib/kubelet/pods/86ec1034-717e-4b83-a8c5-5c6b53987942/volumes" Mar 13 08:34:07 crc kubenswrapper[4876]: I0313 08:34:07.696805 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"cc37175fefd56e14af4bf042273ae967541ee081edd519ea78829a1f09071de2"} Mar 13 08:34:24 crc kubenswrapper[4876]: I0313 08:34:24.169015 4876 scope.go:117] "RemoveContainer" containerID="630cf1e046cb38e12ef0b06bddd7d84d8fafddd15e92b03087a60ceeafce927e" Mar 13 08:35:28 crc kubenswrapper[4876]: I0313 08:35:28.476321 4876 generic.go:334] "Generic (PLEG): container finished" podID="009d8699-15dd-4c89-bd6c-5e7a936e1b59" containerID="583433128f332739101afda35726f7f69c6d0148f2ad7e4a97c110717ce4ce3a" exitCode=0 Mar 13 08:35:28 crc kubenswrapper[4876]: I0313 08:35:28.476451 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"009d8699-15dd-4c89-bd6c-5e7a936e1b59","Type":"ContainerDied","Data":"583433128f332739101afda35726f7f69c6d0148f2ad7e4a97c110717ce4ce3a"} Mar 13 08:35:29 crc kubenswrapper[4876]: I0313 08:35:29.833110 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003266 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ca-certs\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003346 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-workdir\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003368 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ssh-key\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003396 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003443 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-config-data\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003469 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config-secret\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003513 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpj5x\" (UniqueName: \"kubernetes.io/projected/009d8699-15dd-4c89-bd6c-5e7a936e1b59-kube-api-access-fpj5x\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003539 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.003576 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-temporary\") pod \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\" (UID: \"009d8699-15dd-4c89-bd6c-5e7a936e1b59\") " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.005027 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.005179 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-config-data" (OuterVolumeSpecName: "config-data") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.009462 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.010115 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.011474 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/009d8699-15dd-4c89-bd6c-5e7a936e1b59-kube-api-access-fpj5x" (OuterVolumeSpecName: "kube-api-access-fpj5x") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "kube-api-access-fpj5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.033840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.034453 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.035740 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.060089 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "009d8699-15dd-4c89-bd6c-5e7a936e1b59" (UID: "009d8699-15dd-4c89-bd6c-5e7a936e1b59"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105561 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105615 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105634 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpj5x\" (UniqueName: \"kubernetes.io/projected/009d8699-15dd-4c89-bd6c-5e7a936e1b59-kube-api-access-fpj5x\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105648 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/009d8699-15dd-4c89-bd6c-5e7a936e1b59-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105661 4876 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105673 4876 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105684 4876 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/009d8699-15dd-4c89-bd6c-5e7a936e1b59-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105699 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/009d8699-15dd-4c89-bd6c-5e7a936e1b59-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.105746 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.129497 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.208040 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.497760 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"009d8699-15dd-4c89-bd6c-5e7a936e1b59","Type":"ContainerDied","Data":"4c7a7a787683777f0ea7a5daf1786ae418fa67b1459680d04d8d03a9844dccaa"} Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.497805 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c7a7a787683777f0ea7a5daf1786ae418fa67b1459680d04d8d03a9844dccaa" Mar 13 08:35:30 crc kubenswrapper[4876]: I0313 08:35:30.498153 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.603193 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 08:35:41 crc kubenswrapper[4876]: E0313 08:35:41.604416 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009d8699-15dd-4c89-bd6c-5e7a936e1b59" containerName="tempest-tests-tempest-tests-runner" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.604436 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="009d8699-15dd-4c89-bd6c-5e7a936e1b59" containerName="tempest-tests-tempest-tests-runner" Mar 13 08:35:41 crc kubenswrapper[4876]: E0313 08:35:41.604459 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2cee80-a2b8-425d-b58e-54c2710e73a3" containerName="oc" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.604472 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2cee80-a2b8-425d-b58e-54c2710e73a3" containerName="oc" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.604787 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d2cee80-a2b8-425d-b58e-54c2710e73a3" containerName="oc" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.604835 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="009d8699-15dd-4c89-bd6c-5e7a936e1b59" containerName="tempest-tests-tempest-tests-runner" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.605834 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.608878 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nrc9h" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.612411 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.734856 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9ldp\" (UniqueName: \"kubernetes.io/projected/018a8dfd-03af-4cc5-9689-78f7f3472d0d-kube-api-access-p9ldp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"018a8dfd-03af-4cc5-9689-78f7f3472d0d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.734981 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"018a8dfd-03af-4cc5-9689-78f7f3472d0d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.837358 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"018a8dfd-03af-4cc5-9689-78f7f3472d0d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.837641 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9ldp\" (UniqueName: \"kubernetes.io/projected/018a8dfd-03af-4cc5-9689-78f7f3472d0d-kube-api-access-p9ldp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"018a8dfd-03af-4cc5-9689-78f7f3472d0d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.837879 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"018a8dfd-03af-4cc5-9689-78f7f3472d0d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.863482 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9ldp\" (UniqueName: \"kubernetes.io/projected/018a8dfd-03af-4cc5-9689-78f7f3472d0d-kube-api-access-p9ldp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"018a8dfd-03af-4cc5-9689-78f7f3472d0d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.864272 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"018a8dfd-03af-4cc5-9689-78f7f3472d0d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:41 crc kubenswrapper[4876]: I0313 08:35:41.929224 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 08:35:42 crc kubenswrapper[4876]: I0313 08:35:42.377590 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 08:35:42 crc kubenswrapper[4876]: I0313 08:35:42.608647 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"018a8dfd-03af-4cc5-9689-78f7f3472d0d","Type":"ContainerStarted","Data":"198f95400a9a681422164812234afa1cc1dc97f71d8136f57654b077a16db4f7"} Mar 13 08:35:44 crc kubenswrapper[4876]: I0313 08:35:44.626229 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"018a8dfd-03af-4cc5-9689-78f7f3472d0d","Type":"ContainerStarted","Data":"2efbeee1762758fa6eaf82c9daafb1a16d17f613cbc8fce9f652b5db47b7fd20"} Mar 13 08:35:44 crc kubenswrapper[4876]: I0313 08:35:44.654609 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.591714951 podStartE2EDuration="3.654583247s" podCreationTimestamp="2026-03-13 08:35:41 +0000 UTC" firstStartedPulling="2026-03-13 08:35:42.381462391 +0000 UTC m=+3402.052241373" lastFinishedPulling="2026-03-13 08:35:43.444330687 +0000 UTC m=+3403.115109669" observedRunningTime="2026-03-13 08:35:44.643416032 +0000 UTC m=+3404.314195044" watchObservedRunningTime="2026-03-13 08:35:44.654583247 +0000 UTC m=+3404.325362239" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.177807 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556516-cqrfr"] Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.180605 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556516-cqrfr" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.182840 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.183417 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.183620 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.197987 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556516-cqrfr"] Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.200568 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zxrd\" (UniqueName: \"kubernetes.io/projected/5852c991-fc01-46e9-b740-4e716ea321ce-kube-api-access-9zxrd\") pod \"auto-csr-approver-29556516-cqrfr\" (UID: \"5852c991-fc01-46e9-b740-4e716ea321ce\") " pod="openshift-infra/auto-csr-approver-29556516-cqrfr" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.302080 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zxrd\" (UniqueName: \"kubernetes.io/projected/5852c991-fc01-46e9-b740-4e716ea321ce-kube-api-access-9zxrd\") pod \"auto-csr-approver-29556516-cqrfr\" (UID: \"5852c991-fc01-46e9-b740-4e716ea321ce\") " pod="openshift-infra/auto-csr-approver-29556516-cqrfr" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.322449 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zxrd\" (UniqueName: \"kubernetes.io/projected/5852c991-fc01-46e9-b740-4e716ea321ce-kube-api-access-9zxrd\") pod \"auto-csr-approver-29556516-cqrfr\" (UID: \"5852c991-fc01-46e9-b740-4e716ea321ce\") " pod="openshift-infra/auto-csr-approver-29556516-cqrfr" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.508473 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556516-cqrfr" Mar 13 08:36:00 crc kubenswrapper[4876]: I0313 08:36:00.955213 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556516-cqrfr"] Mar 13 08:36:00 crc kubenswrapper[4876]: W0313 08:36:00.978500 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5852c991_fc01_46e9_b740_4e716ea321ce.slice/crio-f236e7d279a800a940e59afa60cd35c870d255832cb4eeae0a197478eb66d1b2 WatchSource:0}: Error finding container f236e7d279a800a940e59afa60cd35c870d255832cb4eeae0a197478eb66d1b2: Status 404 returned error can't find the container with id f236e7d279a800a940e59afa60cd35c870d255832cb4eeae0a197478eb66d1b2 Mar 13 08:36:01 crc kubenswrapper[4876]: I0313 08:36:01.802920 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556516-cqrfr" event={"ID":"5852c991-fc01-46e9-b740-4e716ea321ce","Type":"ContainerStarted","Data":"f236e7d279a800a940e59afa60cd35c870d255832cb4eeae0a197478eb66d1b2"} Mar 13 08:36:02 crc kubenswrapper[4876]: I0313 08:36:02.815997 4876 generic.go:334] "Generic (PLEG): container finished" podID="5852c991-fc01-46e9-b740-4e716ea321ce" containerID="93b509967beab31a37b2e2ceaf3ae117a2d2372bc888e1ed9b4ede0106cbcf82" exitCode=0 Mar 13 08:36:02 crc kubenswrapper[4876]: I0313 08:36:02.816067 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556516-cqrfr" event={"ID":"5852c991-fc01-46e9-b740-4e716ea321ce","Type":"ContainerDied","Data":"93b509967beab31a37b2e2ceaf3ae117a2d2372bc888e1ed9b4ede0106cbcf82"} Mar 13 08:36:04 crc kubenswrapper[4876]: I0313 08:36:04.222530 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556516-cqrfr" Mar 13 08:36:04 crc kubenswrapper[4876]: I0313 08:36:04.390950 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zxrd\" (UniqueName: \"kubernetes.io/projected/5852c991-fc01-46e9-b740-4e716ea321ce-kube-api-access-9zxrd\") pod \"5852c991-fc01-46e9-b740-4e716ea321ce\" (UID: \"5852c991-fc01-46e9-b740-4e716ea321ce\") " Mar 13 08:36:04 crc kubenswrapper[4876]: I0313 08:36:04.397382 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5852c991-fc01-46e9-b740-4e716ea321ce-kube-api-access-9zxrd" (OuterVolumeSpecName: "kube-api-access-9zxrd") pod "5852c991-fc01-46e9-b740-4e716ea321ce" (UID: "5852c991-fc01-46e9-b740-4e716ea321ce"). InnerVolumeSpecName "kube-api-access-9zxrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:36:04 crc kubenswrapper[4876]: I0313 08:36:04.492807 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zxrd\" (UniqueName: \"kubernetes.io/projected/5852c991-fc01-46e9-b740-4e716ea321ce-kube-api-access-9zxrd\") on node \"crc\" DevicePath \"\"" Mar 13 08:36:04 crc kubenswrapper[4876]: I0313 08:36:04.847190 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556516-cqrfr" event={"ID":"5852c991-fc01-46e9-b740-4e716ea321ce","Type":"ContainerDied","Data":"f236e7d279a800a940e59afa60cd35c870d255832cb4eeae0a197478eb66d1b2"} Mar 13 08:36:04 crc kubenswrapper[4876]: I0313 08:36:04.847654 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f236e7d279a800a940e59afa60cd35c870d255832cb4eeae0a197478eb66d1b2" Mar 13 08:36:04 crc kubenswrapper[4876]: I0313 08:36:04.847298 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556516-cqrfr" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.300248 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556510-96s8s"] Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.308116 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556510-96s8s"] Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.820145 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mczk2/must-gather-4mvdh"] Mar 13 08:36:05 crc kubenswrapper[4876]: E0313 08:36:05.820677 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5852c991-fc01-46e9-b740-4e716ea321ce" containerName="oc" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.820705 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5852c991-fc01-46e9-b740-4e716ea321ce" containerName="oc" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.820946 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5852c991-fc01-46e9-b740-4e716ea321ce" containerName="oc" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.823121 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.829741 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mczk2"/"default-dockercfg-t97hn" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.833167 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mczk2"/"openshift-service-ca.crt" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.833207 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mczk2"/"kube-root-ca.crt" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.845510 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mczk2/must-gather-4mvdh"] Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.920430 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6ba56888-06b5-406f-b194-0a6e11a72675-must-gather-output\") pod \"must-gather-4mvdh\" (UID: \"6ba56888-06b5-406f-b194-0a6e11a72675\") " pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:36:05 crc kubenswrapper[4876]: I0313 08:36:05.920679 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7wm7\" (UniqueName: \"kubernetes.io/projected/6ba56888-06b5-406f-b194-0a6e11a72675-kube-api-access-q7wm7\") pod \"must-gather-4mvdh\" (UID: \"6ba56888-06b5-406f-b194-0a6e11a72675\") " pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:36:06 crc kubenswrapper[4876]: I0313 08:36:06.022448 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7wm7\" (UniqueName: \"kubernetes.io/projected/6ba56888-06b5-406f-b194-0a6e11a72675-kube-api-access-q7wm7\") pod \"must-gather-4mvdh\" (UID: \"6ba56888-06b5-406f-b194-0a6e11a72675\") " pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:36:06 crc kubenswrapper[4876]: I0313 08:36:06.022598 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6ba56888-06b5-406f-b194-0a6e11a72675-must-gather-output\") pod \"must-gather-4mvdh\" (UID: \"6ba56888-06b5-406f-b194-0a6e11a72675\") " pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:36:06 crc kubenswrapper[4876]: I0313 08:36:06.023221 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6ba56888-06b5-406f-b194-0a6e11a72675-must-gather-output\") pod \"must-gather-4mvdh\" (UID: \"6ba56888-06b5-406f-b194-0a6e11a72675\") " pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:36:06 crc kubenswrapper[4876]: I0313 08:36:06.047416 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7wm7\" (UniqueName: \"kubernetes.io/projected/6ba56888-06b5-406f-b194-0a6e11a72675-kube-api-access-q7wm7\") pod \"must-gather-4mvdh\" (UID: \"6ba56888-06b5-406f-b194-0a6e11a72675\") " pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:36:06 crc kubenswrapper[4876]: I0313 08:36:06.150674 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:36:06 crc kubenswrapper[4876]: I0313 08:36:06.706473 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mczk2/must-gather-4mvdh"] Mar 13 08:36:06 crc kubenswrapper[4876]: I0313 08:36:06.874914 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/must-gather-4mvdh" event={"ID":"6ba56888-06b5-406f-b194-0a6e11a72675","Type":"ContainerStarted","Data":"b0d9da16a66f495f5646128ea1f5e84861e850a1ad609effdaab1e6dba0eb838"} Mar 13 08:36:07 crc kubenswrapper[4876]: I0313 08:36:07.044411 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cab4aa9-1c76-48b3-b8cc-14fc1c07820b" path="/var/lib/kubelet/pods/9cab4aa9-1c76-48b3-b8cc-14fc1c07820b/volumes" Mar 13 08:36:12 crc kubenswrapper[4876]: I0313 08:36:12.945537 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/must-gather-4mvdh" event={"ID":"6ba56888-06b5-406f-b194-0a6e11a72675","Type":"ContainerStarted","Data":"d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069"} Mar 13 08:36:13 crc kubenswrapper[4876]: I0313 08:36:13.958668 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/must-gather-4mvdh" event={"ID":"6ba56888-06b5-406f-b194-0a6e11a72675","Type":"ContainerStarted","Data":"72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d"} Mar 13 08:36:13 crc kubenswrapper[4876]: I0313 08:36:13.977422 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mczk2/must-gather-4mvdh" podStartSLOduration=3.14408617 podStartE2EDuration="8.977370101s" podCreationTimestamp="2026-03-13 08:36:05 +0000 UTC" firstStartedPulling="2026-03-13 08:36:06.778936333 +0000 UTC m=+3426.449715315" lastFinishedPulling="2026-03-13 08:36:12.612220244 +0000 UTC m=+3432.282999246" observedRunningTime="2026-03-13 08:36:13.976673231 +0000 UTC m=+3433.647452213" watchObservedRunningTime="2026-03-13 08:36:13.977370101 +0000 UTC m=+3433.648149113" Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.491159 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mczk2/crc-debug-cxc55"] Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.493074 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.680742 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f7992084-2410-4342-93d1-c1c057505a3f-host\") pod \"crc-debug-cxc55\" (UID: \"f7992084-2410-4342-93d1-c1c057505a3f\") " pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.680892 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx47b\" (UniqueName: \"kubernetes.io/projected/f7992084-2410-4342-93d1-c1c057505a3f-kube-api-access-nx47b\") pod \"crc-debug-cxc55\" (UID: \"f7992084-2410-4342-93d1-c1c057505a3f\") " pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.782278 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx47b\" (UniqueName: \"kubernetes.io/projected/f7992084-2410-4342-93d1-c1c057505a3f-kube-api-access-nx47b\") pod \"crc-debug-cxc55\" (UID: \"f7992084-2410-4342-93d1-c1c057505a3f\") " pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.782429 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f7992084-2410-4342-93d1-c1c057505a3f-host\") pod \"crc-debug-cxc55\" (UID: \"f7992084-2410-4342-93d1-c1c057505a3f\") " pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.782568 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f7992084-2410-4342-93d1-c1c057505a3f-host\") pod \"crc-debug-cxc55\" (UID: \"f7992084-2410-4342-93d1-c1c057505a3f\") " pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.803518 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx47b\" (UniqueName: \"kubernetes.io/projected/f7992084-2410-4342-93d1-c1c057505a3f-kube-api-access-nx47b\") pod \"crc-debug-cxc55\" (UID: \"f7992084-2410-4342-93d1-c1c057505a3f\") " pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.811834 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:36:17 crc kubenswrapper[4876]: W0313 08:36:17.860127 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7992084_2410_4342_93d1_c1c057505a3f.slice/crio-9d02e41d0bcbff59dd88cee1052206bf13cebbaae251ed9973f6fe836a020006 WatchSource:0}: Error finding container 9d02e41d0bcbff59dd88cee1052206bf13cebbaae251ed9973f6fe836a020006: Status 404 returned error can't find the container with id 9d02e41d0bcbff59dd88cee1052206bf13cebbaae251ed9973f6fe836a020006 Mar 13 08:36:17 crc kubenswrapper[4876]: I0313 08:36:17.999069 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/crc-debug-cxc55" event={"ID":"f7992084-2410-4342-93d1-c1c057505a3f","Type":"ContainerStarted","Data":"9d02e41d0bcbff59dd88cee1052206bf13cebbaae251ed9973f6fe836a020006"} Mar 13 08:36:24 crc kubenswrapper[4876]: I0313 08:36:24.253817 4876 scope.go:117] "RemoveContainer" containerID="a77192805bafccf47da7f79e035324dc4fdd928380259b1d34da6a9f1f339732" Mar 13 08:36:24 crc kubenswrapper[4876]: I0313 08:36:24.607912 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:36:24 crc kubenswrapper[4876]: I0313 08:36:24.608572 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:36:30 crc kubenswrapper[4876]: I0313 08:36:30.148607 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/crc-debug-cxc55" event={"ID":"f7992084-2410-4342-93d1-c1c057505a3f","Type":"ContainerStarted","Data":"cc5382bbe00aef499349eb64c022c168d548f3d969f164f811a26639dc3bc9c8"} Mar 13 08:36:30 crc kubenswrapper[4876]: I0313 08:36:30.168192 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mczk2/crc-debug-cxc55" podStartSLOduration=1.7746367680000001 podStartE2EDuration="13.168159184s" podCreationTimestamp="2026-03-13 08:36:17 +0000 UTC" firstStartedPulling="2026-03-13 08:36:17.863254316 +0000 UTC m=+3437.534033298" lastFinishedPulling="2026-03-13 08:36:29.256776732 +0000 UTC m=+3448.927555714" observedRunningTime="2026-03-13 08:36:30.161622978 +0000 UTC m=+3449.832401970" watchObservedRunningTime="2026-03-13 08:36:30.168159184 +0000 UTC m=+3449.838938166" Mar 13 08:36:54 crc kubenswrapper[4876]: I0313 08:36:54.607513 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:36:54 crc kubenswrapper[4876]: I0313 08:36:54.608831 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:37:08 crc kubenswrapper[4876]: I0313 08:37:08.491268 4876 generic.go:334] "Generic (PLEG): container finished" podID="f7992084-2410-4342-93d1-c1c057505a3f" containerID="cc5382bbe00aef499349eb64c022c168d548f3d969f164f811a26639dc3bc9c8" exitCode=0 Mar 13 08:37:08 crc kubenswrapper[4876]: I0313 08:37:08.491314 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/crc-debug-cxc55" event={"ID":"f7992084-2410-4342-93d1-c1c057505a3f","Type":"ContainerDied","Data":"cc5382bbe00aef499349eb64c022c168d548f3d969f164f811a26639dc3bc9c8"} Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.618634 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.659495 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mczk2/crc-debug-cxc55"] Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.668567 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mczk2/crc-debug-cxc55"] Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.722269 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f7992084-2410-4342-93d1-c1c057505a3f-host\") pod \"f7992084-2410-4342-93d1-c1c057505a3f\" (UID: \"f7992084-2410-4342-93d1-c1c057505a3f\") " Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.722370 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7992084-2410-4342-93d1-c1c057505a3f-host" (OuterVolumeSpecName: "host") pod "f7992084-2410-4342-93d1-c1c057505a3f" (UID: "f7992084-2410-4342-93d1-c1c057505a3f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.722496 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx47b\" (UniqueName: \"kubernetes.io/projected/f7992084-2410-4342-93d1-c1c057505a3f-kube-api-access-nx47b\") pod \"f7992084-2410-4342-93d1-c1c057505a3f\" (UID: \"f7992084-2410-4342-93d1-c1c057505a3f\") " Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.722984 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f7992084-2410-4342-93d1-c1c057505a3f-host\") on node \"crc\" DevicePath \"\"" Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.730025 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7992084-2410-4342-93d1-c1c057505a3f-kube-api-access-nx47b" (OuterVolumeSpecName: "kube-api-access-nx47b") pod "f7992084-2410-4342-93d1-c1c057505a3f" (UID: "f7992084-2410-4342-93d1-c1c057505a3f"). InnerVolumeSpecName "kube-api-access-nx47b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:37:09 crc kubenswrapper[4876]: I0313 08:37:09.824779 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx47b\" (UniqueName: \"kubernetes.io/projected/f7992084-2410-4342-93d1-c1c057505a3f-kube-api-access-nx47b\") on node \"crc\" DevicePath \"\"" Mar 13 08:37:10 crc kubenswrapper[4876]: I0313 08:37:10.510091 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d02e41d0bcbff59dd88cee1052206bf13cebbaae251ed9973f6fe836a020006" Mar 13 08:37:10 crc kubenswrapper[4876]: I0313 08:37:10.510519 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-cxc55" Mar 13 08:37:10 crc kubenswrapper[4876]: I0313 08:37:10.863808 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mczk2/crc-debug-q8vp5"] Mar 13 08:37:10 crc kubenswrapper[4876]: E0313 08:37:10.864477 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7992084-2410-4342-93d1-c1c057505a3f" containerName="container-00" Mar 13 08:37:10 crc kubenswrapper[4876]: I0313 08:37:10.864499 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7992084-2410-4342-93d1-c1c057505a3f" containerName="container-00" Mar 13 08:37:10 crc kubenswrapper[4876]: I0313 08:37:10.864827 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7992084-2410-4342-93d1-c1c057505a3f" containerName="container-00" Mar 13 08:37:10 crc kubenswrapper[4876]: I0313 08:37:10.865829 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:10 crc kubenswrapper[4876]: I0313 08:37:10.945044 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea5f298b-38df-402c-a630-a79682d7c54a-host\") pod \"crc-debug-q8vp5\" (UID: \"ea5f298b-38df-402c-a630-a79682d7c54a\") " pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:10 crc kubenswrapper[4876]: I0313 08:37:10.945130 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cd6g\" (UniqueName: \"kubernetes.io/projected/ea5f298b-38df-402c-a630-a79682d7c54a-kube-api-access-2cd6g\") pod \"crc-debug-q8vp5\" (UID: \"ea5f298b-38df-402c-a630-a79682d7c54a\") " pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.045872 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7992084-2410-4342-93d1-c1c057505a3f" path="/var/lib/kubelet/pods/f7992084-2410-4342-93d1-c1c057505a3f/volumes" Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.047432 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea5f298b-38df-402c-a630-a79682d7c54a-host\") pod \"crc-debug-q8vp5\" (UID: \"ea5f298b-38df-402c-a630-a79682d7c54a\") " pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.047500 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cd6g\" (UniqueName: \"kubernetes.io/projected/ea5f298b-38df-402c-a630-a79682d7c54a-kube-api-access-2cd6g\") pod \"crc-debug-q8vp5\" (UID: \"ea5f298b-38df-402c-a630-a79682d7c54a\") " pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.047574 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea5f298b-38df-402c-a630-a79682d7c54a-host\") pod \"crc-debug-q8vp5\" (UID: \"ea5f298b-38df-402c-a630-a79682d7c54a\") " pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.072031 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cd6g\" (UniqueName: \"kubernetes.io/projected/ea5f298b-38df-402c-a630-a79682d7c54a-kube-api-access-2cd6g\") pod \"crc-debug-q8vp5\" (UID: \"ea5f298b-38df-402c-a630-a79682d7c54a\") " pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.186639 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.518274 4876 generic.go:334] "Generic (PLEG): container finished" podID="ea5f298b-38df-402c-a630-a79682d7c54a" containerID="f5b7b7398afa8b53e90855ff2d7d446690a3229da40041a9e5797dfd075025f6" exitCode=0 Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.518606 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/crc-debug-q8vp5" event={"ID":"ea5f298b-38df-402c-a630-a79682d7c54a","Type":"ContainerDied","Data":"f5b7b7398afa8b53e90855ff2d7d446690a3229da40041a9e5797dfd075025f6"} Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.518631 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/crc-debug-q8vp5" event={"ID":"ea5f298b-38df-402c-a630-a79682d7c54a","Type":"ContainerStarted","Data":"ef9cc565e5bf86abb7fbf67b37c801f3a69793bca38011296e1edef96bee398f"} Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.893462 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mczk2/crc-debug-q8vp5"] Mar 13 08:37:11 crc kubenswrapper[4876]: I0313 08:37:11.901225 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mczk2/crc-debug-q8vp5"] Mar 13 08:37:12 crc kubenswrapper[4876]: I0313 08:37:12.621746 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:12 crc kubenswrapper[4876]: I0313 08:37:12.670381 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea5f298b-38df-402c-a630-a79682d7c54a-host\") pod \"ea5f298b-38df-402c-a630-a79682d7c54a\" (UID: \"ea5f298b-38df-402c-a630-a79682d7c54a\") " Mar 13 08:37:12 crc kubenswrapper[4876]: I0313 08:37:12.670462 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cd6g\" (UniqueName: \"kubernetes.io/projected/ea5f298b-38df-402c-a630-a79682d7c54a-kube-api-access-2cd6g\") pod \"ea5f298b-38df-402c-a630-a79682d7c54a\" (UID: \"ea5f298b-38df-402c-a630-a79682d7c54a\") " Mar 13 08:37:12 crc kubenswrapper[4876]: I0313 08:37:12.670561 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea5f298b-38df-402c-a630-a79682d7c54a-host" (OuterVolumeSpecName: "host") pod "ea5f298b-38df-402c-a630-a79682d7c54a" (UID: "ea5f298b-38df-402c-a630-a79682d7c54a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 08:37:12 crc kubenswrapper[4876]: I0313 08:37:12.671024 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea5f298b-38df-402c-a630-a79682d7c54a-host\") on node \"crc\" DevicePath \"\"" Mar 13 08:37:12 crc kubenswrapper[4876]: I0313 08:37:12.688408 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea5f298b-38df-402c-a630-a79682d7c54a-kube-api-access-2cd6g" (OuterVolumeSpecName: "kube-api-access-2cd6g") pod "ea5f298b-38df-402c-a630-a79682d7c54a" (UID: "ea5f298b-38df-402c-a630-a79682d7c54a"). InnerVolumeSpecName "kube-api-access-2cd6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:37:12 crc kubenswrapper[4876]: I0313 08:37:12.773451 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cd6g\" (UniqueName: \"kubernetes.io/projected/ea5f298b-38df-402c-a630-a79682d7c54a-kube-api-access-2cd6g\") on node \"crc\" DevicePath \"\"" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.048965 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea5f298b-38df-402c-a630-a79682d7c54a" path="/var/lib/kubelet/pods/ea5f298b-38df-402c-a630-a79682d7c54a/volumes" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.096986 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mczk2/crc-debug-wcj7c"] Mar 13 08:37:13 crc kubenswrapper[4876]: E0313 08:37:13.097580 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5f298b-38df-402c-a630-a79682d7c54a" containerName="container-00" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.097606 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5f298b-38df-402c-a630-a79682d7c54a" containerName="container-00" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.097833 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea5f298b-38df-402c-a630-a79682d7c54a" containerName="container-00" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.105566 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.314285 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-host\") pod \"crc-debug-wcj7c\" (UID: \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\") " pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.318513 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrszn\" (UniqueName: \"kubernetes.io/projected/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-kube-api-access-nrszn\") pod \"crc-debug-wcj7c\" (UID: \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\") " pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.424303 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-host\") pod \"crc-debug-wcj7c\" (UID: \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\") " pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.424320 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-host\") pod \"crc-debug-wcj7c\" (UID: \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\") " pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.424443 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrszn\" (UniqueName: \"kubernetes.io/projected/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-kube-api-access-nrszn\") pod \"crc-debug-wcj7c\" (UID: \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\") " pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.442884 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrszn\" (UniqueName: \"kubernetes.io/projected/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-kube-api-access-nrszn\") pod \"crc-debug-wcj7c\" (UID: \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\") " pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.537315 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-q8vp5" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.537944 4876 scope.go:117] "RemoveContainer" containerID="f5b7b7398afa8b53e90855ff2d7d446690a3229da40041a9e5797dfd075025f6" Mar 13 08:37:13 crc kubenswrapper[4876]: I0313 08:37:13.724098 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:13 crc kubenswrapper[4876]: W0313 08:37:13.759316 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa5d23e2_8946_4adb_94fb_c5cc67d57cf3.slice/crio-912a1a786552a4b35be59a3eb42a9e8ee3e39c538a5bf369016f69e824c7b9c3 WatchSource:0}: Error finding container 912a1a786552a4b35be59a3eb42a9e8ee3e39c538a5bf369016f69e824c7b9c3: Status 404 returned error can't find the container with id 912a1a786552a4b35be59a3eb42a9e8ee3e39c538a5bf369016f69e824c7b9c3 Mar 13 08:37:14 crc kubenswrapper[4876]: I0313 08:37:14.549968 4876 generic.go:334] "Generic (PLEG): container finished" podID="aa5d23e2-8946-4adb-94fb-c5cc67d57cf3" containerID="83f3e3f86bf127fa9374eb6af308ec1ce09fa9111bf905c35515b3571de246e2" exitCode=0 Mar 13 08:37:14 crc kubenswrapper[4876]: I0313 08:37:14.550081 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/crc-debug-wcj7c" event={"ID":"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3","Type":"ContainerDied","Data":"83f3e3f86bf127fa9374eb6af308ec1ce09fa9111bf905c35515b3571de246e2"} Mar 13 08:37:14 crc kubenswrapper[4876]: I0313 08:37:14.550367 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/crc-debug-wcj7c" event={"ID":"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3","Type":"ContainerStarted","Data":"912a1a786552a4b35be59a3eb42a9e8ee3e39c538a5bf369016f69e824c7b9c3"} Mar 13 08:37:14 crc kubenswrapper[4876]: I0313 08:37:14.594531 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mczk2/crc-debug-wcj7c"] Mar 13 08:37:14 crc kubenswrapper[4876]: I0313 08:37:14.604537 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mczk2/crc-debug-wcj7c"] Mar 13 08:37:15 crc kubenswrapper[4876]: I0313 08:37:15.652884 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:15 crc kubenswrapper[4876]: I0313 08:37:15.764298 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-host\") pod \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\" (UID: \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\") " Mar 13 08:37:15 crc kubenswrapper[4876]: I0313 08:37:15.764453 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-host" (OuterVolumeSpecName: "host") pod "aa5d23e2-8946-4adb-94fb-c5cc67d57cf3" (UID: "aa5d23e2-8946-4adb-94fb-c5cc67d57cf3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 08:37:15 crc kubenswrapper[4876]: I0313 08:37:15.764470 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrszn\" (UniqueName: \"kubernetes.io/projected/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-kube-api-access-nrszn\") pod \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\" (UID: \"aa5d23e2-8946-4adb-94fb-c5cc67d57cf3\") " Mar 13 08:37:15 crc kubenswrapper[4876]: I0313 08:37:15.764991 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-host\") on node \"crc\" DevicePath \"\"" Mar 13 08:37:15 crc kubenswrapper[4876]: I0313 08:37:15.769873 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-kube-api-access-nrszn" (OuterVolumeSpecName: "kube-api-access-nrszn") pod "aa5d23e2-8946-4adb-94fb-c5cc67d57cf3" (UID: "aa5d23e2-8946-4adb-94fb-c5cc67d57cf3"). InnerVolumeSpecName "kube-api-access-nrszn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:37:15 crc kubenswrapper[4876]: I0313 08:37:15.867018 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrszn\" (UniqueName: \"kubernetes.io/projected/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3-kube-api-access-nrszn\") on node \"crc\" DevicePath \"\"" Mar 13 08:37:16 crc kubenswrapper[4876]: I0313 08:37:16.569645 4876 scope.go:117] "RemoveContainer" containerID="83f3e3f86bf127fa9374eb6af308ec1ce09fa9111bf905c35515b3571de246e2" Mar 13 08:37:16 crc kubenswrapper[4876]: I0313 08:37:16.569931 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/crc-debug-wcj7c" Mar 13 08:37:17 crc kubenswrapper[4876]: I0313 08:37:17.047760 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa5d23e2-8946-4adb-94fb-c5cc67d57cf3" path="/var/lib/kubelet/pods/aa5d23e2-8946-4adb-94fb-c5cc67d57cf3/volumes" Mar 13 08:37:24 crc kubenswrapper[4876]: I0313 08:37:24.607172 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:37:24 crc kubenswrapper[4876]: I0313 08:37:24.607737 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:37:24 crc kubenswrapper[4876]: I0313 08:37:24.607779 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:37:24 crc kubenswrapper[4876]: I0313 08:37:24.608581 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cc37175fefd56e14af4bf042273ae967541ee081edd519ea78829a1f09071de2"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:37:24 crc kubenswrapper[4876]: I0313 08:37:24.608633 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://cc37175fefd56e14af4bf042273ae967541ee081edd519ea78829a1f09071de2" gracePeriod=600 Mar 13 08:37:25 crc kubenswrapper[4876]: I0313 08:37:25.649864 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="cc37175fefd56e14af4bf042273ae967541ee081edd519ea78829a1f09071de2" exitCode=0 Mar 13 08:37:25 crc kubenswrapper[4876]: I0313 08:37:25.649982 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"cc37175fefd56e14af4bf042273ae967541ee081edd519ea78829a1f09071de2"} Mar 13 08:37:25 crc kubenswrapper[4876]: I0313 08:37:25.650479 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55"} Mar 13 08:37:25 crc kubenswrapper[4876]: I0313 08:37:25.650502 4876 scope.go:117] "RemoveContainer" containerID="7e4bcfef38d2981d47ec733bf76201c725d4b78d180e633f855982b924dd4b0b" Mar 13 08:37:30 crc kubenswrapper[4876]: I0313 08:37:30.961994 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8b86bd47d-2hv8s_65a5d87e-b9ef-4be3-ade6-46fa03282d7d/barbican-api/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.111026 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8b86bd47d-2hv8s_65a5d87e-b9ef-4be3-ade6-46fa03282d7d/barbican-api-log/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.172536 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5864c7c4b6-qqb8c_79f28ab7-204b-4e1c-be09-9e48b9ed5a75/barbican-keystone-listener/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.192338 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5864c7c4b6-qqb8c_79f28ab7-204b-4e1c-be09-9e48b9ed5a75/barbican-keystone-listener-log/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.319444 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5467d6f7c5-djg95_94630a52-4ec8-4ba9-966b-334ec46e9d25/barbican-worker/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.344633 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5467d6f7c5-djg95_94630a52-4ec8-4ba9-966b-334ec46e9d25/barbican-worker-log/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.550513 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z_0808c5a4-53f5-43e5-af39-bf78530e1d71/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.823064 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_27e28286-b51c-43cf-acbb-d612cae68530/ceilometer-notification-agent/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.829216 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_27e28286-b51c-43cf-acbb-d612cae68530/ceilometer-central-agent/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.872585 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_27e28286-b51c-43cf-acbb-d612cae68530/proxy-httpd/0.log" Mar 13 08:37:31 crc kubenswrapper[4876]: I0313 08:37:31.897481 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_27e28286-b51c-43cf-acbb-d612cae68530/sg-core/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.026594 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_19442bc3-f0a0-4077-8b11-823aeb39dc3f/cinder-api/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.074515 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_19442bc3-f0a0-4077-8b11-823aeb39dc3f/cinder-api-log/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.298471 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c3cb708-df89-4a10-ba84-b0c55efd2b61/probe/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.299323 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c3cb708-df89-4a10-ba84-b0c55efd2b61/cinder-scheduler/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.377868 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5_5bd6dd59-7427-4e76-8570-0a334476406c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.501720 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4_5268b132-b205-4741-8680-781f73f3209a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.583365 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-vvnsb_6de4174c-e8e2-4058-a35b-c277eced33f9/init/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.860394 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-vvnsb_6de4174c-e8e2-4058-a35b-c277eced33f9/init/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.905428 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-fwnht_0de3f5bc-8761-491e-9a81-9a03e796ca1b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:32 crc kubenswrapper[4876]: I0313 08:37:32.924887 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-vvnsb_6de4174c-e8e2-4058-a35b-c277eced33f9/dnsmasq-dns/0.log" Mar 13 08:37:33 crc kubenswrapper[4876]: I0313 08:37:33.103679 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e4def3c5-fa14-4aed-94c0-3374552b56cf/glance-httpd/0.log" Mar 13 08:37:33 crc kubenswrapper[4876]: I0313 08:37:33.104176 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e4def3c5-fa14-4aed-94c0-3374552b56cf/glance-log/0.log" Mar 13 08:37:33 crc kubenswrapper[4876]: I0313 08:37:33.255685 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_dd9df755-7613-4abe-a9fc-661e6667eb63/glance-httpd/0.log" Mar 13 08:37:33 crc kubenswrapper[4876]: I0313 08:37:33.287973 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_dd9df755-7613-4abe-a9fc-661e6667eb63/glance-log/0.log" Mar 13 08:37:33 crc kubenswrapper[4876]: I0313 08:37:33.540416 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-764bff8b56-4l8lg_70b62196-7476-4c85-a2d7-2299da069989/horizon/0.log" Mar 13 08:37:33 crc kubenswrapper[4876]: I0313 08:37:33.683809 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xdclw_92d53dff-55bb-4953-96a9-647552c446e0/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:33 crc kubenswrapper[4876]: I0313 08:37:33.777958 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4m5cw_fb7b397a-4ce3-4a83-b236-5baac12e2a15/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:33 crc kubenswrapper[4876]: I0313 08:37:33.799620 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-764bff8b56-4l8lg_70b62196-7476-4c85-a2d7-2299da069989/horizon-log/0.log" Mar 13 08:37:34 crc kubenswrapper[4876]: I0313 08:37:34.071146 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29556481-xrkdk_a91beee0-58e6-4d98-8000-e4e6bbee83a8/keystone-cron/0.log" Mar 13 08:37:34 crc kubenswrapper[4876]: I0313 08:37:34.079287 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-556bf8599b-szn97_b9195731-b1ca-4c6c-90f3-54549fa3a813/keystone-api/0.log" Mar 13 08:37:34 crc kubenswrapper[4876]: I0313 08:37:34.296642 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_ff27b563-91dc-4be1-b0b8-98955850d2dd/kube-state-metrics/0.log" Mar 13 08:37:34 crc kubenswrapper[4876]: I0313 08:37:34.300229 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k_240f0880-695b-44b4-ae52-2beacd34f2d0/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:34 crc kubenswrapper[4876]: I0313 08:37:34.713176 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5884ccb8d9-4wjtm_cfccefa5-a3f2-4513-a79b-372dd903ac2d/neutron-api/0.log" Mar 13 08:37:34 crc kubenswrapper[4876]: I0313 08:37:34.779265 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5884ccb8d9-4wjtm_cfccefa5-a3f2-4513-a79b-372dd903ac2d/neutron-httpd/0.log" Mar 13 08:37:34 crc kubenswrapper[4876]: I0313 08:37:34.954426 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg_97951031-bd23-462e-83a8-be7068e1dfc8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:35 crc kubenswrapper[4876]: I0313 08:37:35.496875 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cfd0230f-78c9-4fa3-a450-889afd1fb40a/nova-api-log/0.log" Mar 13 08:37:35 crc kubenswrapper[4876]: I0313 08:37:35.530273 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a77871ec-0f52-4d09-a3e2-664d14c9adbf/nova-cell0-conductor-conductor/0.log" Mar 13 08:37:35 crc kubenswrapper[4876]: I0313 08:37:35.794681 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cfd0230f-78c9-4fa3-a450-889afd1fb40a/nova-api-api/0.log" Mar 13 08:37:35 crc kubenswrapper[4876]: I0313 08:37:35.805910 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83/nova-cell1-conductor-conductor/0.log" Mar 13 08:37:35 crc kubenswrapper[4876]: I0313 08:37:35.831116 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9b56d8da-0553-46dd-b2ab-26df263e1885/nova-cell1-novncproxy-novncproxy/0.log" Mar 13 08:37:36 crc kubenswrapper[4876]: I0313 08:37:36.137274 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-2xhz4_7b648455-e602-439e-b806-743233acb285/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:36 crc kubenswrapper[4876]: I0313 08:37:36.185802 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_33193662-0ae1-4e6d-bcb8-d093ee42b647/nova-metadata-log/0.log" Mar 13 08:37:36 crc kubenswrapper[4876]: I0313 08:37:36.479674 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e8c36ecf-187d-4388-8d6f-626fe6470eb0/mysql-bootstrap/0.log" Mar 13 08:37:36 crc kubenswrapper[4876]: I0313 08:37:36.532407 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d51bcb66-a417-4369-94df-a59aedd82d1c/nova-scheduler-scheduler/0.log" Mar 13 08:37:36 crc kubenswrapper[4876]: I0313 08:37:36.656403 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e8c36ecf-187d-4388-8d6f-626fe6470eb0/mysql-bootstrap/0.log" Mar 13 08:37:36 crc kubenswrapper[4876]: I0313 08:37:36.705971 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e8c36ecf-187d-4388-8d6f-626fe6470eb0/galera/0.log" Mar 13 08:37:36 crc kubenswrapper[4876]: I0313 08:37:36.848864 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7e32caff-0034-47f2-aa5b-917607dd44b6/mysql-bootstrap/0.log" Mar 13 08:37:37 crc kubenswrapper[4876]: I0313 08:37:37.201031 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7e32caff-0034-47f2-aa5b-917607dd44b6/galera/0.log" Mar 13 08:37:37 crc kubenswrapper[4876]: I0313 08:37:37.216040 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7e32caff-0034-47f2-aa5b-917607dd44b6/mysql-bootstrap/0.log" Mar 13 08:37:37 crc kubenswrapper[4876]: I0313 08:37:37.382826 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_02e40021-e864-4c53-9522-40fe5719abde/openstackclient/0.log" Mar 13 08:37:37 crc kubenswrapper[4876]: I0313 08:37:37.382953 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_33193662-0ae1-4e6d-bcb8-d093ee42b647/nova-metadata-metadata/0.log" Mar 13 08:37:37 crc kubenswrapper[4876]: I0313 08:37:37.529044 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rgh2p_d0dbe7f9-9cce-4b19-86d6-1b445c996d95/openstack-network-exporter/0.log" Mar 13 08:37:37 crc kubenswrapper[4876]: I0313 08:37:37.611628 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tb4jh_23f73ce7-85a0-410c-a3f3-f1df187b509e/ovsdb-server-init/0.log" Mar 13 08:37:37 crc kubenswrapper[4876]: I0313 08:37:37.911906 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tb4jh_23f73ce7-85a0-410c-a3f3-f1df187b509e/ovsdb-server-init/0.log" Mar 13 08:37:37 crc kubenswrapper[4876]: I0313 08:37:37.931063 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tb4jh_23f73ce7-85a0-410c-a3f3-f1df187b509e/ovsdb-server/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.018346 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tb4jh_23f73ce7-85a0-410c-a3f3-f1df187b509e/ovs-vswitchd/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.205609 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xrn6k_3b55481b-c5a5-4602-8f66-c3b002134edb/ovn-controller/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.263186 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hrnbj_848b0256-b349-49c3-a007-20c0da303ad8/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.381999 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_76257d47-12f0-4e22-a3d7-856367d8a08b/openstack-network-exporter/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.481357 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_76257d47-12f0-4e22-a3d7-856367d8a08b/ovn-northd/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.529712 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_88e4b876-8c78-4d88-945b-bd10555503d2/openstack-network-exporter/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.614037 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_88e4b876-8c78-4d88-945b-bd10555503d2/ovsdbserver-nb/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.684775 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37b39bdf-9f1c-4f03-88e5-6abd3949fc7c/openstack-network-exporter/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.784788 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37b39bdf-9f1c-4f03-88e5-6abd3949fc7c/ovsdbserver-sb/0.log" Mar 13 08:37:38 crc kubenswrapper[4876]: I0313 08:37:38.958218 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6dbd6d4d74-7zn9n_9ab4e2aa-c297-4979-9322-78f14c7b36f8/placement-api/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.041736 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6dbd6d4d74-7zn9n_9ab4e2aa-c297-4979-9322-78f14c7b36f8/placement-log/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.151824 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147c88b9-db25-435e-9575-edb4efa342e7/setup-container/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.381613 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147c88b9-db25-435e-9575-edb4efa342e7/rabbitmq/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.412525 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147c88b9-db25-435e-9575-edb4efa342e7/setup-container/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.444130 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_79aa8084-23ef-4730-948d-1b0d6ca25251/setup-container/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.643723 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_79aa8084-23ef-4730-948d-1b0d6ca25251/setup-container/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.644000 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf_8caf73f8-76f3-4fa6-bfbc-9977576c8d18/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.673447 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_79aa8084-23ef-4730-948d-1b0d6ca25251/rabbitmq/0.log" Mar 13 08:37:39 crc kubenswrapper[4876]: I0313 08:37:39.807844 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-cm4qh_1f20780c-14f1-4165-95a5-15b604018e4b/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.022610 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg_61be635e-fed6-4b19-9095-cbaa3720b693/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.117995 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-hj2tl_f87ce9a3-0a04-4b72-bd4f-88a5538c99c1/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.238762 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-6q2l4_ca0807c6-78af-45cf-994d-85563d8089ac/ssh-known-hosts-edpm-deployment/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.476122 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b9c5d6d4c-7lxwx_62fd9eb3-6af0-4bc9-bc57-65c8b03eb467/proxy-server/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.547965 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b9c5d6d4c-7lxwx_62fd9eb3-6af0-4bc9-bc57-65c8b03eb467/proxy-httpd/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.581225 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-pbnh2_e5294658-a1f0-4340-9d61-c8008a37b5e9/swift-ring-rebalance/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.784211 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/account-reaper/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.834099 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/account-auditor/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.872065 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/account-replicator/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.973180 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/account-server/0.log" Mar 13 08:37:40 crc kubenswrapper[4876]: I0313 08:37:40.991941 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/container-auditor/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.087552 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/container-replicator/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.095780 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/container-server/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.170581 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/container-updater/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.278606 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-auditor/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.306258 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-expirer/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.371100 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-replicator/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.372719 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-server/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.506025 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-updater/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.540792 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/rsync/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.568229 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/swift-recon-cron/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.790436 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_009d8699-15dd-4c89-bd6c-5e7a936e1b59/tempest-tests-tempest-tests-runner/0.log" Mar 13 08:37:41 crc kubenswrapper[4876]: I0313 08:37:41.811368 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx_18be12a3-56e2-4b39-8c8f-784a7cafed2b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:42 crc kubenswrapper[4876]: I0313 08:37:42.005333 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_018a8dfd-03af-4cc5-9689-78f7f3472d0d/test-operator-logs-container/0.log" Mar 13 08:37:42 crc kubenswrapper[4876]: I0313 08:37:42.047637 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qljdp_5b3fa1e6-19bc-4852-bf70-69051aa084d6/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:37:50 crc kubenswrapper[4876]: I0313 08:37:50.882956 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e221cd11-36ee-4c17-a26d-e95b75ed5682/memcached/0.log" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.546736 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5xbzh"] Mar 13 08:37:51 crc kubenswrapper[4876]: E0313 08:37:51.547284 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5d23e2-8946-4adb-94fb-c5cc67d57cf3" containerName="container-00" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.547307 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5d23e2-8946-4adb-94fb-c5cc67d57cf3" containerName="container-00" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.547574 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5d23e2-8946-4adb-94fb-c5cc67d57cf3" containerName="container-00" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.548960 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.573512 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xbzh"] Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.729672 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-utilities\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.729881 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-catalog-content\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.730210 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xm45\" (UniqueName: \"kubernetes.io/projected/34e361c7-ee1a-458b-9de4-6d27ae1a657a-kube-api-access-8xm45\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.831935 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xm45\" (UniqueName: \"kubernetes.io/projected/34e361c7-ee1a-458b-9de4-6d27ae1a657a-kube-api-access-8xm45\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.832045 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-utilities\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.832122 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-catalog-content\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.832733 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-catalog-content\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.832845 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-utilities\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.852388 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xm45\" (UniqueName: \"kubernetes.io/projected/34e361c7-ee1a-458b-9de4-6d27ae1a657a-kube-api-access-8xm45\") pod \"redhat-marketplace-5xbzh\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:51 crc kubenswrapper[4876]: I0313 08:37:51.870715 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:37:52 crc kubenswrapper[4876]: I0313 08:37:52.398643 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xbzh"] Mar 13 08:37:52 crc kubenswrapper[4876]: I0313 08:37:52.892985 4876 generic.go:334] "Generic (PLEG): container finished" podID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerID="a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4" exitCode=0 Mar 13 08:37:52 crc kubenswrapper[4876]: I0313 08:37:52.893044 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xbzh" event={"ID":"34e361c7-ee1a-458b-9de4-6d27ae1a657a","Type":"ContainerDied","Data":"a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4"} Mar 13 08:37:52 crc kubenswrapper[4876]: I0313 08:37:52.893318 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xbzh" event={"ID":"34e361c7-ee1a-458b-9de4-6d27ae1a657a","Type":"ContainerStarted","Data":"05d429876d353627ed09ea498ed9c063105858a0b8e5a1e7bc9f7ea389d4b332"} Mar 13 08:37:52 crc kubenswrapper[4876]: I0313 08:37:52.894992 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:37:53 crc kubenswrapper[4876]: I0313 08:37:53.904114 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xbzh" event={"ID":"34e361c7-ee1a-458b-9de4-6d27ae1a657a","Type":"ContainerStarted","Data":"d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc"} Mar 13 08:37:54 crc kubenswrapper[4876]: I0313 08:37:54.925647 4876 generic.go:334] "Generic (PLEG): container finished" podID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerID="d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc" exitCode=0 Mar 13 08:37:54 crc kubenswrapper[4876]: I0313 08:37:54.925708 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xbzh" event={"ID":"34e361c7-ee1a-458b-9de4-6d27ae1a657a","Type":"ContainerDied","Data":"d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc"} Mar 13 08:37:55 crc kubenswrapper[4876]: I0313 08:37:55.934736 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xbzh" event={"ID":"34e361c7-ee1a-458b-9de4-6d27ae1a657a","Type":"ContainerStarted","Data":"0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906"} Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.154317 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5xbzh" podStartSLOduration=6.760680503 podStartE2EDuration="9.154228763s" podCreationTimestamp="2026-03-13 08:37:51 +0000 UTC" firstStartedPulling="2026-03-13 08:37:52.894722622 +0000 UTC m=+3532.565501604" lastFinishedPulling="2026-03-13 08:37:55.288270882 +0000 UTC m=+3534.959049864" observedRunningTime="2026-03-13 08:37:55.955218276 +0000 UTC m=+3535.625997278" watchObservedRunningTime="2026-03-13 08:38:00.154228763 +0000 UTC m=+3539.825007765" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.157067 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556518-5jtkd"] Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.159173 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556518-5jtkd" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.162159 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.162629 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.164034 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.185009 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556518-5jtkd"] Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.307490 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqs4k\" (UniqueName: \"kubernetes.io/projected/a38b573d-799d-4983-b44b-5f2ec7867aa3-kube-api-access-kqs4k\") pod \"auto-csr-approver-29556518-5jtkd\" (UID: \"a38b573d-799d-4983-b44b-5f2ec7867aa3\") " pod="openshift-infra/auto-csr-approver-29556518-5jtkd" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.409227 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqs4k\" (UniqueName: \"kubernetes.io/projected/a38b573d-799d-4983-b44b-5f2ec7867aa3-kube-api-access-kqs4k\") pod \"auto-csr-approver-29556518-5jtkd\" (UID: \"a38b573d-799d-4983-b44b-5f2ec7867aa3\") " pod="openshift-infra/auto-csr-approver-29556518-5jtkd" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.428342 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqs4k\" (UniqueName: \"kubernetes.io/projected/a38b573d-799d-4983-b44b-5f2ec7867aa3-kube-api-access-kqs4k\") pod \"auto-csr-approver-29556518-5jtkd\" (UID: \"a38b573d-799d-4983-b44b-5f2ec7867aa3\") " pod="openshift-infra/auto-csr-approver-29556518-5jtkd" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.482459 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556518-5jtkd" Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.912928 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556518-5jtkd"] Mar 13 08:38:00 crc kubenswrapper[4876]: I0313 08:38:00.980833 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556518-5jtkd" event={"ID":"a38b573d-799d-4983-b44b-5f2ec7867aa3","Type":"ContainerStarted","Data":"14488324cd070b4ad9ba6f4013326825e9acd42c3f1eb931419d9a902e8b65ff"} Mar 13 08:38:01 crc kubenswrapper[4876]: I0313 08:38:01.871679 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:38:01 crc kubenswrapper[4876]: I0313 08:38:01.871745 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:38:01 crc kubenswrapper[4876]: I0313 08:38:01.921501 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:38:02 crc kubenswrapper[4876]: I0313 08:38:02.030907 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:38:02 crc kubenswrapper[4876]: I0313 08:38:02.174716 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xbzh"] Mar 13 08:38:02 crc kubenswrapper[4876]: I0313 08:38:02.998482 4876 generic.go:334] "Generic (PLEG): container finished" podID="a38b573d-799d-4983-b44b-5f2ec7867aa3" containerID="a15c8d5e1499b9ea2bd1483974ca56f67bbf4fceb6d7b9b9476926d64acae7df" exitCode=0 Mar 13 08:38:02 crc kubenswrapper[4876]: I0313 08:38:02.998560 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556518-5jtkd" event={"ID":"a38b573d-799d-4983-b44b-5f2ec7867aa3","Type":"ContainerDied","Data":"a15c8d5e1499b9ea2bd1483974ca56f67bbf4fceb6d7b9b9476926d64acae7df"} Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.008479 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5xbzh" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerName="registry-server" containerID="cri-o://0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906" gracePeriod=2 Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.390293 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556518-5jtkd" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.488114 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqs4k\" (UniqueName: \"kubernetes.io/projected/a38b573d-799d-4983-b44b-5f2ec7867aa3-kube-api-access-kqs4k\") pod \"a38b573d-799d-4983-b44b-5f2ec7867aa3\" (UID: \"a38b573d-799d-4983-b44b-5f2ec7867aa3\") " Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.499027 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a38b573d-799d-4983-b44b-5f2ec7867aa3-kube-api-access-kqs4k" (OuterVolumeSpecName: "kube-api-access-kqs4k") pod "a38b573d-799d-4983-b44b-5f2ec7867aa3" (UID: "a38b573d-799d-4983-b44b-5f2ec7867aa3"). InnerVolumeSpecName "kube-api-access-kqs4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.539676 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.594616 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqs4k\" (UniqueName: \"kubernetes.io/projected/a38b573d-799d-4983-b44b-5f2ec7867aa3-kube-api-access-kqs4k\") on node \"crc\" DevicePath \"\"" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.696358 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xm45\" (UniqueName: \"kubernetes.io/projected/34e361c7-ee1a-458b-9de4-6d27ae1a657a-kube-api-access-8xm45\") pod \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.696575 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-utilities\") pod \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.696617 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-catalog-content\") pod \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\" (UID: \"34e361c7-ee1a-458b-9de4-6d27ae1a657a\") " Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.697551 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-utilities" (OuterVolumeSpecName: "utilities") pod "34e361c7-ee1a-458b-9de4-6d27ae1a657a" (UID: "34e361c7-ee1a-458b-9de4-6d27ae1a657a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.700551 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34e361c7-ee1a-458b-9de4-6d27ae1a657a-kube-api-access-8xm45" (OuterVolumeSpecName: "kube-api-access-8xm45") pod "34e361c7-ee1a-458b-9de4-6d27ae1a657a" (UID: "34e361c7-ee1a-458b-9de4-6d27ae1a657a"). InnerVolumeSpecName "kube-api-access-8xm45". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.724530 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34e361c7-ee1a-458b-9de4-6d27ae1a657a" (UID: "34e361c7-ee1a-458b-9de4-6d27ae1a657a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.798697 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xm45\" (UniqueName: \"kubernetes.io/projected/34e361c7-ee1a-458b-9de4-6d27ae1a657a-kube-api-access-8xm45\") on node \"crc\" DevicePath \"\"" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.798730 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:38:04 crc kubenswrapper[4876]: I0313 08:38:04.798740 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34e361c7-ee1a-458b-9de4-6d27ae1a657a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.018647 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556518-5jtkd" event={"ID":"a38b573d-799d-4983-b44b-5f2ec7867aa3","Type":"ContainerDied","Data":"14488324cd070b4ad9ba6f4013326825e9acd42c3f1eb931419d9a902e8b65ff"} Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.018951 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14488324cd070b4ad9ba6f4013326825e9acd42c3f1eb931419d9a902e8b65ff" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.018688 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556518-5jtkd" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.022114 4876 generic.go:334] "Generic (PLEG): container finished" podID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerID="0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906" exitCode=0 Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.022158 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xbzh" event={"ID":"34e361c7-ee1a-458b-9de4-6d27ae1a657a","Type":"ContainerDied","Data":"0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906"} Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.022189 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xbzh" event={"ID":"34e361c7-ee1a-458b-9de4-6d27ae1a657a","Type":"ContainerDied","Data":"05d429876d353627ed09ea498ed9c063105858a0b8e5a1e7bc9f7ea389d4b332"} Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.022210 4876 scope.go:117] "RemoveContainer" containerID="0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.022275 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xbzh" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.061569 4876 scope.go:117] "RemoveContainer" containerID="d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.089201 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xbzh"] Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.099587 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xbzh"] Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.113021 4876 scope.go:117] "RemoveContainer" containerID="a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.179216 4876 scope.go:117] "RemoveContainer" containerID="0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906" Mar 13 08:38:05 crc kubenswrapper[4876]: E0313 08:38:05.179696 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906\": container with ID starting with 0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906 not found: ID does not exist" containerID="0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.179735 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906"} err="failed to get container status \"0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906\": rpc error: code = NotFound desc = could not find container \"0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906\": container with ID starting with 0548f041159df4abd9eee11d2ffd09525d67bf597282c4af22205e5afe79f906 not found: ID does not exist" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.179770 4876 scope.go:117] "RemoveContainer" containerID="d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc" Mar 13 08:38:05 crc kubenswrapper[4876]: E0313 08:38:05.180286 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc\": container with ID starting with d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc not found: ID does not exist" containerID="d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.180324 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc"} err="failed to get container status \"d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc\": rpc error: code = NotFound desc = could not find container \"d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc\": container with ID starting with d67c5fd70005976add7852c5fcb4893f7db8a09e05da6689f92820703886b1cc not found: ID does not exist" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.180346 4876 scope.go:117] "RemoveContainer" containerID="a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4" Mar 13 08:38:05 crc kubenswrapper[4876]: E0313 08:38:05.180592 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4\": container with ID starting with a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4 not found: ID does not exist" containerID="a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.180618 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4"} err="failed to get container status \"a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4\": rpc error: code = NotFound desc = could not find container \"a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4\": container with ID starting with a9339f494544db22c30f7d9a5c26ed63c8aff7915ca7531f6ea956a5f6aa8db4 not found: ID does not exist" Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.467131 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556512-z7ps6"] Mar 13 08:38:05 crc kubenswrapper[4876]: I0313 08:38:05.497764 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556512-z7ps6"] Mar 13 08:38:07 crc kubenswrapper[4876]: I0313 08:38:07.044188 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" path="/var/lib/kubelet/pods/34e361c7-ee1a-458b-9de4-6d27ae1a657a/volumes" Mar 13 08:38:07 crc kubenswrapper[4876]: I0313 08:38:07.045334 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e09b27f-c606-4be9-83c9-152efce10f10" path="/var/lib/kubelet/pods/7e09b27f-c606-4be9-83c9-152efce10f10/volumes" Mar 13 08:38:07 crc kubenswrapper[4876]: I0313 08:38:07.502012 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/util/0.log" Mar 13 08:38:07 crc kubenswrapper[4876]: I0313 08:38:07.790193 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/util/0.log" Mar 13 08:38:07 crc kubenswrapper[4876]: I0313 08:38:07.794608 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/pull/0.log" Mar 13 08:38:07 crc kubenswrapper[4876]: I0313 08:38:07.843486 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/pull/0.log" Mar 13 08:38:08 crc kubenswrapper[4876]: I0313 08:38:08.094169 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/util/0.log" Mar 13 08:38:08 crc kubenswrapper[4876]: I0313 08:38:08.113864 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/pull/0.log" Mar 13 08:38:08 crc kubenswrapper[4876]: I0313 08:38:08.113908 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/extract/0.log" Mar 13 08:38:08 crc kubenswrapper[4876]: I0313 08:38:08.520738 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-897gd_68223df4-94ca-47fe-afbc-93c77725ba2a/manager/0.log" Mar 13 08:38:08 crc kubenswrapper[4876]: I0313 08:38:08.853732 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-zlt8c_151ed757-e99e-4e7e-bfaa-7dec0ef8b038/manager/0.log" Mar 13 08:38:09 crc kubenswrapper[4876]: I0313 08:38:09.041982 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-jcjfd_8c04b15f-3ca3-4875-9ea4-45244808af5f/manager/0.log" Mar 13 08:38:09 crc kubenswrapper[4876]: I0313 08:38:09.291640 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-c6bfj_ff48d75e-19f2-4a48-bcd6-3855b494fdf1/manager/0.log" Mar 13 08:38:09 crc kubenswrapper[4876]: I0313 08:38:09.720779 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-w45l7_8d21a99d-11a5-4b5f-a44d-3f9cce84b82a/manager/0.log" Mar 13 08:38:09 crc kubenswrapper[4876]: I0313 08:38:09.777141 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6bbb499bbc-n5stb_3ff4d1b1-aae8-4900-a317-164ea72a78b2/manager/0.log" Mar 13 08:38:09 crc kubenswrapper[4876]: I0313 08:38:09.925457 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5995f4446f-lcwhv_722c5d70-ed42-4b75-8993-3b3643d44d1a/manager/0.log" Mar 13 08:38:10 crc kubenswrapper[4876]: I0313 08:38:10.100567 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-kmhm8_33e3356e-590a-4b31-990f-4ecda38efb0a/manager/0.log" Mar 13 08:38:10 crc kubenswrapper[4876]: I0313 08:38:10.157555 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-68f45f9d9f-bp9bt_98ac6099-e731-43c8-8537-8095f1bfce09/manager/0.log" Mar 13 08:38:10 crc kubenswrapper[4876]: I0313 08:38:10.420734 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-658d4cdd5-v7kpw_6aeda87b-2ede-4dcb-9234-bfc8837f4ea8/manager/0.log" Mar 13 08:38:10 crc kubenswrapper[4876]: I0313 08:38:10.577373 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-lnb48_4284fe2d-d1bb-49f6-b555-0f46419ad11b/manager/0.log" Mar 13 08:38:10 crc kubenswrapper[4876]: I0313 08:38:10.764452 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-569cc54c5-pfswh_ef934388-2da5-47fe-8679-ec520946de1b/manager/0.log" Mar 13 08:38:10 crc kubenswrapper[4876]: I0313 08:38:10.876024 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-m2xqm_3de3f00b-74de-4328-86e9-9f86c7277deb/manager/0.log" Mar 13 08:38:11 crc kubenswrapper[4876]: I0313 08:38:11.046428 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q_256daf30-6602-4a5d-8a4b-0f96fb5450b2/manager/0.log" Mar 13 08:38:11 crc kubenswrapper[4876]: I0313 08:38:11.398588 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-578cd8b599-6vlp7_995e15f2-9906-40df-a498-832366249bf9/operator/0.log" Mar 13 08:38:11 crc kubenswrapper[4876]: I0313 08:38:11.759641 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tfbgk_ec8ee7ae-49f7-473f-b432-9c2395e93115/registry-server/0.log" Mar 13 08:38:11 crc kubenswrapper[4876]: I0313 08:38:11.962984 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-d5d5r_dacb8b62-d4f3-43af-88b2-26fbc725f50f/manager/0.log" Mar 13 08:38:12 crc kubenswrapper[4876]: I0313 08:38:12.224461 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-m7zdz_a9a4fe04-37f1-47ad-ab32-e983f13df515/manager/0.log" Mar 13 08:38:12 crc kubenswrapper[4876]: I0313 08:38:12.242186 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-tzshq_8a6e1029-8d6f-4698-b92b-d0cd17bda62d/operator/0.log" Mar 13 08:38:12 crc kubenswrapper[4876]: I0313 08:38:12.419361 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-677c674df7-hgmsc_e7f3a695-6649-482f-94d9-238966224ba8/manager/0.log" Mar 13 08:38:12 crc kubenswrapper[4876]: I0313 08:38:12.709931 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cd66dbd4b-jt5rp_a314a819-a02b-4e9f-a996-58bed1d2e6be/manager/0.log" Mar 13 08:38:12 crc kubenswrapper[4876]: I0313 08:38:12.762837 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-jztw5_78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e/manager/0.log" Mar 13 08:38:12 crc kubenswrapper[4876]: I0313 08:38:12.947782 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6dd88c6f67-cm65k_3aefe2ee-3d59-46e9-9a6e-4f576566a49c/manager/0.log" Mar 13 08:38:13 crc kubenswrapper[4876]: I0313 08:38:13.117996 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74f6d57fdc-mfkh6_bad784ad-2cd8-4f50-832f-0e57d06249d6/manager/0.log" Mar 13 08:38:13 crc kubenswrapper[4876]: I0313 08:38:13.923891 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-6slg2_4cfaf185-ed2d-4c36-9d38-fcf4e74905ec/manager/0.log" Mar 13 08:38:29 crc kubenswrapper[4876]: I0313 08:38:29.288712 4876 scope.go:117] "RemoveContainer" containerID="0cefe81b31844959742d7080e6ac8ccfc509616720396f9ce49c94513fad867a" Mar 13 08:38:32 crc kubenswrapper[4876]: I0313 08:38:32.192177 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-hln48_8c6d92f4-c514-48e4-b1da-61546f07d171/control-plane-machine-set-operator/0.log" Mar 13 08:38:32 crc kubenswrapper[4876]: I0313 08:38:32.375382 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xvwj2_418bec94-3753-4d76-b2a8-24b04f39c820/kube-rbac-proxy/0.log" Mar 13 08:38:32 crc kubenswrapper[4876]: I0313 08:38:32.442820 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xvwj2_418bec94-3753-4d76-b2a8-24b04f39c820/machine-api-operator/0.log" Mar 13 08:38:45 crc kubenswrapper[4876]: I0313 08:38:45.984223 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-552r7_bbd6238b-a711-47f6-bf54-ae2b35e8c79d/cert-manager-controller/0.log" Mar 13 08:38:46 crc kubenswrapper[4876]: I0313 08:38:46.156855 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-2c8kw_037a0101-dff7-49a6-9974-a5b88c2cb7f1/cert-manager-cainjector/0.log" Mar 13 08:38:46 crc kubenswrapper[4876]: I0313 08:38:46.213315 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-jm7bb_d0aa3593-da37-4da6-ae1f-890cb901c029/cert-manager-webhook/0.log" Mar 13 08:38:58 crc kubenswrapper[4876]: I0313 08:38:58.584073 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-62blj_be89d0d3-8483-4439-b346-c50bd885b15c/nmstate-console-plugin/0.log" Mar 13 08:38:58 crc kubenswrapper[4876]: I0313 08:38:58.771596 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-v9cql_6aa5c930-3567-401a-830f-c30ca123b57a/nmstate-handler/0.log" Mar 13 08:38:58 crc kubenswrapper[4876]: I0313 08:38:58.876771 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-d2mlg_8f70b2a8-97e0-476b-91d2-e5c386d8a14f/kube-rbac-proxy/0.log" Mar 13 08:38:58 crc kubenswrapper[4876]: I0313 08:38:58.921956 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-d2mlg_8f70b2a8-97e0-476b-91d2-e5c386d8a14f/nmstate-metrics/0.log" Mar 13 08:38:59 crc kubenswrapper[4876]: I0313 08:38:59.078055 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-5gv4s_54e0cdfc-3234-4763-bc49-6a6e7a537c74/nmstate-operator/0.log" Mar 13 08:38:59 crc kubenswrapper[4876]: I0313 08:38:59.115380 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-74bqr_e506e0e1-25c2-4be3-ad95-5f9029a960e3/nmstate-webhook/0.log" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.007324 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rk2h8"] Mar 13 08:39:22 crc kubenswrapper[4876]: E0313 08:39:22.008609 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerName="registry-server" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.008635 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerName="registry-server" Mar 13 08:39:22 crc kubenswrapper[4876]: E0313 08:39:22.008670 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38b573d-799d-4983-b44b-5f2ec7867aa3" containerName="oc" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.008681 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38b573d-799d-4983-b44b-5f2ec7867aa3" containerName="oc" Mar 13 08:39:22 crc kubenswrapper[4876]: E0313 08:39:22.008703 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerName="extract-content" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.008715 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerName="extract-content" Mar 13 08:39:22 crc kubenswrapper[4876]: E0313 08:39:22.008766 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerName="extract-utilities" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.008777 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerName="extract-utilities" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.009074 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a38b573d-799d-4983-b44b-5f2ec7867aa3" containerName="oc" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.009097 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="34e361c7-ee1a-458b-9de4-6d27ae1a657a" containerName="registry-server" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.011777 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.018499 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rk2h8"] Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.063261 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-utilities\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.063344 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrjb6\" (UniqueName: \"kubernetes.io/projected/91661619-2ece-4221-a278-5d1945d57a57-kube-api-access-hrjb6\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.063530 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-catalog-content\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.165191 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrjb6\" (UniqueName: \"kubernetes.io/projected/91661619-2ece-4221-a278-5d1945d57a57-kube-api-access-hrjb6\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.165582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-catalog-content\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.165745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-utilities\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.166098 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-catalog-content\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.166214 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-utilities\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.205294 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrjb6\" (UniqueName: \"kubernetes.io/projected/91661619-2ece-4221-a278-5d1945d57a57-kube-api-access-hrjb6\") pod \"redhat-operators-rk2h8\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.345786 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:22 crc kubenswrapper[4876]: I0313 08:39:22.870358 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rk2h8"] Mar 13 08:39:23 crc kubenswrapper[4876]: I0313 08:39:23.742581 4876 generic.go:334] "Generic (PLEG): container finished" podID="91661619-2ece-4221-a278-5d1945d57a57" containerID="f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689" exitCode=0 Mar 13 08:39:23 crc kubenswrapper[4876]: I0313 08:39:23.742664 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk2h8" event={"ID":"91661619-2ece-4221-a278-5d1945d57a57","Type":"ContainerDied","Data":"f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689"} Mar 13 08:39:23 crc kubenswrapper[4876]: I0313 08:39:23.743099 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk2h8" event={"ID":"91661619-2ece-4221-a278-5d1945d57a57","Type":"ContainerStarted","Data":"f14f257f3b26d1f4ae5de9726c1684c7dcaf6403254e3dec0eb65ce19c4162ba"} Mar 13 08:39:24 crc kubenswrapper[4876]: I0313 08:39:24.607285 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:39:24 crc kubenswrapper[4876]: I0313 08:39:24.607620 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:39:24 crc kubenswrapper[4876]: I0313 08:39:24.770618 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk2h8" event={"ID":"91661619-2ece-4221-a278-5d1945d57a57","Type":"ContainerStarted","Data":"d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21"} Mar 13 08:39:25 crc kubenswrapper[4876]: I0313 08:39:25.779795 4876 generic.go:334] "Generic (PLEG): container finished" podID="91661619-2ece-4221-a278-5d1945d57a57" containerID="d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21" exitCode=0 Mar 13 08:39:25 crc kubenswrapper[4876]: I0313 08:39:25.779887 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk2h8" event={"ID":"91661619-2ece-4221-a278-5d1945d57a57","Type":"ContainerDied","Data":"d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21"} Mar 13 08:39:26 crc kubenswrapper[4876]: I0313 08:39:26.790528 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk2h8" event={"ID":"91661619-2ece-4221-a278-5d1945d57a57","Type":"ContainerStarted","Data":"ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01"} Mar 13 08:39:26 crc kubenswrapper[4876]: I0313 08:39:26.807438 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rk2h8" podStartSLOduration=3.340636383 podStartE2EDuration="5.80741928s" podCreationTimestamp="2026-03-13 08:39:21 +0000 UTC" firstStartedPulling="2026-03-13 08:39:23.74419296 +0000 UTC m=+3623.414971932" lastFinishedPulling="2026-03-13 08:39:26.210975847 +0000 UTC m=+3625.881754829" observedRunningTime="2026-03-13 08:39:26.804598873 +0000 UTC m=+3626.475377855" watchObservedRunningTime="2026-03-13 08:39:26.80741928 +0000 UTC m=+3626.478198262" Mar 13 08:39:27 crc kubenswrapper[4876]: I0313 08:39:27.487480 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-bmq26_97a22a43-697e-4c6a-b800-489008955cbe/kube-rbac-proxy/0.log" Mar 13 08:39:27 crc kubenswrapper[4876]: I0313 08:39:27.670732 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-bmq26_97a22a43-697e-4c6a-b800-489008955cbe/controller/0.log" Mar 13 08:39:27 crc kubenswrapper[4876]: I0313 08:39:27.821755 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-frr-files/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.036883 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-frr-files/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.102300 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-metrics/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.128928 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-reloader/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.163026 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-reloader/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.435014 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-reloader/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.447024 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-frr-files/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.455208 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-metrics/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.476977 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-metrics/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.690487 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-frr-files/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.698816 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-metrics/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.717824 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/controller/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.744620 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-reloader/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.920155 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/frr-metrics/0.log" Mar 13 08:39:28 crc kubenswrapper[4876]: I0313 08:39:28.969320 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/kube-rbac-proxy/0.log" Mar 13 08:39:29 crc kubenswrapper[4876]: I0313 08:39:29.090600 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/kube-rbac-proxy-frr/0.log" Mar 13 08:39:29 crc kubenswrapper[4876]: I0313 08:39:29.301318 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-6vvzq_91b6f5ca-3921-446a-a57a-de264ca63118/frr-k8s-webhook-server/0.log" Mar 13 08:39:29 crc kubenswrapper[4876]: I0313 08:39:29.338643 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/reloader/0.log" Mar 13 08:39:29 crc kubenswrapper[4876]: I0313 08:39:29.829192 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5f79f8644-vxb58_122ad6c8-564c-4aff-8dc5-6515b0e9b5cb/manager/0.log" Mar 13 08:39:29 crc kubenswrapper[4876]: I0313 08:39:29.913784 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-679dfbf9b7-f5wpn_e918dae3-24c9-4e8f-8f89-f4070580eeff/webhook-server/0.log" Mar 13 08:39:30 crc kubenswrapper[4876]: I0313 08:39:30.115464 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nb74f_65146c95-889a-4030-b7d3-3a1810b158de/kube-rbac-proxy/0.log" Mar 13 08:39:30 crc kubenswrapper[4876]: I0313 08:39:30.774820 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nb74f_65146c95-889a-4030-b7d3-3a1810b158de/speaker/0.log" Mar 13 08:39:30 crc kubenswrapper[4876]: I0313 08:39:30.832499 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/frr/0.log" Mar 13 08:39:32 crc kubenswrapper[4876]: I0313 08:39:32.346035 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:32 crc kubenswrapper[4876]: I0313 08:39:32.346417 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:39:33 crc kubenswrapper[4876]: I0313 08:39:33.400741 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rk2h8" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="registry-server" probeResult="failure" output=< Mar 13 08:39:33 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:39:33 crc kubenswrapper[4876]: > Mar 13 08:39:43 crc kubenswrapper[4876]: I0313 08:39:43.398748 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rk2h8" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="registry-server" probeResult="failure" output=< Mar 13 08:39:43 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:39:43 crc kubenswrapper[4876]: > Mar 13 08:39:43 crc kubenswrapper[4876]: I0313 08:39:43.648370 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/util/0.log" Mar 13 08:39:43 crc kubenswrapper[4876]: I0313 08:39:43.802496 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/util/0.log" Mar 13 08:39:43 crc kubenswrapper[4876]: I0313 08:39:43.884770 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/pull/0.log" Mar 13 08:39:43 crc kubenswrapper[4876]: I0313 08:39:43.891540 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/pull/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.057409 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/util/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.136332 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/pull/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.158565 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/extract/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.444683 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/util/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.573408 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/pull/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.592020 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/util/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.645125 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/pull/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.761922 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/util/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.782994 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/pull/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.807497 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/extract/0.log" Mar 13 08:39:44 crc kubenswrapper[4876]: I0313 08:39:44.942689 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-utilities/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.147607 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-content/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.196312 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-utilities/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.244154 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-content/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.360225 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-utilities/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.421478 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-content/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.619800 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-utilities/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.794524 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-utilities/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.859297 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-content/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.913910 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/registry-server/0.log" Mar 13 08:39:45 crc kubenswrapper[4876]: I0313 08:39:45.941979 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-content/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.149148 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-content/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.161930 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-utilities/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.359357 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gzmpk_a0adab6c-7942-4756-829f-3b7a6569f8cb/marketplace-operator/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.427201 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-utilities/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.667845 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-utilities/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.668996 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/registry-server/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.741180 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-content/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.760445 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-content/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.965027 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-content/0.log" Mar 13 08:39:46 crc kubenswrapper[4876]: I0313 08:39:46.966814 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-utilities/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.104551 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/registry-server/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.204091 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rk2h8_91661619-2ece-4221-a278-5d1945d57a57/extract-utilities/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.382156 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rk2h8_91661619-2ece-4221-a278-5d1945d57a57/extract-utilities/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.384565 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rk2h8_91661619-2ece-4221-a278-5d1945d57a57/extract-content/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.405159 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rk2h8_91661619-2ece-4221-a278-5d1945d57a57/extract-content/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.602562 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rk2h8_91661619-2ece-4221-a278-5d1945d57a57/extract-utilities/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.645674 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rk2h8_91661619-2ece-4221-a278-5d1945d57a57/extract-content/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.672168 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rk2h8_91661619-2ece-4221-a278-5d1945d57a57/registry-server/0.log" Mar 13 08:39:47 crc kubenswrapper[4876]: I0313 08:39:47.799108 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-utilities/0.log" Mar 13 08:39:48 crc kubenswrapper[4876]: I0313 08:39:48.119358 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-content/0.log" Mar 13 08:39:48 crc kubenswrapper[4876]: I0313 08:39:48.151608 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-utilities/0.log" Mar 13 08:39:48 crc kubenswrapper[4876]: I0313 08:39:48.198812 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-content/0.log" Mar 13 08:39:48 crc kubenswrapper[4876]: I0313 08:39:48.376995 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-utilities/0.log" Mar 13 08:39:48 crc kubenswrapper[4876]: I0313 08:39:48.407011 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-content/0.log" Mar 13 08:39:48 crc kubenswrapper[4876]: I0313 08:39:48.921654 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/registry-server/0.log" Mar 13 08:39:53 crc kubenswrapper[4876]: I0313 08:39:53.395446 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rk2h8" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="registry-server" probeResult="failure" output=< Mar 13 08:39:53 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:39:53 crc kubenswrapper[4876]: > Mar 13 08:39:54 crc kubenswrapper[4876]: I0313 08:39:54.607606 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:39:54 crc kubenswrapper[4876]: I0313 08:39:54.607662 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.146340 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556520-nm5cz"] Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.148069 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556520-nm5cz" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.150453 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.150507 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.151215 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.157373 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556520-nm5cz"] Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.261152 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr797\" (UniqueName: \"kubernetes.io/projected/e7631eb4-cbcc-4df8-a4a8-46f00224c41e-kube-api-access-zr797\") pod \"auto-csr-approver-29556520-nm5cz\" (UID: \"e7631eb4-cbcc-4df8-a4a8-46f00224c41e\") " pod="openshift-infra/auto-csr-approver-29556520-nm5cz" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.363529 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr797\" (UniqueName: \"kubernetes.io/projected/e7631eb4-cbcc-4df8-a4a8-46f00224c41e-kube-api-access-zr797\") pod \"auto-csr-approver-29556520-nm5cz\" (UID: \"e7631eb4-cbcc-4df8-a4a8-46f00224c41e\") " pod="openshift-infra/auto-csr-approver-29556520-nm5cz" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.385730 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr797\" (UniqueName: \"kubernetes.io/projected/e7631eb4-cbcc-4df8-a4a8-46f00224c41e-kube-api-access-zr797\") pod \"auto-csr-approver-29556520-nm5cz\" (UID: \"e7631eb4-cbcc-4df8-a4a8-46f00224c41e\") " pod="openshift-infra/auto-csr-approver-29556520-nm5cz" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.475549 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556520-nm5cz" Mar 13 08:40:00 crc kubenswrapper[4876]: I0313 08:40:00.927802 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556520-nm5cz"] Mar 13 08:40:01 crc kubenswrapper[4876]: I0313 08:40:01.092078 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556520-nm5cz" event={"ID":"e7631eb4-cbcc-4df8-a4a8-46f00224c41e","Type":"ContainerStarted","Data":"62f3aafecac63aaa65f8debb53e9542a2771061dc97cf992b4028db056946244"} Mar 13 08:40:02 crc kubenswrapper[4876]: I0313 08:40:02.397202 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:40:02 crc kubenswrapper[4876]: I0313 08:40:02.448739 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:40:02 crc kubenswrapper[4876]: I0313 08:40:02.654477 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rk2h8"] Mar 13 08:40:03 crc kubenswrapper[4876]: I0313 08:40:03.109541 4876 generic.go:334] "Generic (PLEG): container finished" podID="e7631eb4-cbcc-4df8-a4a8-46f00224c41e" containerID="2cf7f22e2f77c14f865113d9307e85633f40a1b93af8c479e357b60cf251b911" exitCode=0 Mar 13 08:40:03 crc kubenswrapper[4876]: I0313 08:40:03.109861 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556520-nm5cz" event={"ID":"e7631eb4-cbcc-4df8-a4a8-46f00224c41e","Type":"ContainerDied","Data":"2cf7f22e2f77c14f865113d9307e85633f40a1b93af8c479e357b60cf251b911"} Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.118252 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rk2h8" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="registry-server" containerID="cri-o://ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01" gracePeriod=2 Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.638079 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556520-nm5cz" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.672873 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr797\" (UniqueName: \"kubernetes.io/projected/e7631eb4-cbcc-4df8-a4a8-46f00224c41e-kube-api-access-zr797\") pod \"e7631eb4-cbcc-4df8-a4a8-46f00224c41e\" (UID: \"e7631eb4-cbcc-4df8-a4a8-46f00224c41e\") " Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.697393 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7631eb4-cbcc-4df8-a4a8-46f00224c41e-kube-api-access-zr797" (OuterVolumeSpecName: "kube-api-access-zr797") pod "e7631eb4-cbcc-4df8-a4a8-46f00224c41e" (UID: "e7631eb4-cbcc-4df8-a4a8-46f00224c41e"). InnerVolumeSpecName "kube-api-access-zr797". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.749431 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.775715 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-utilities\") pod \"91661619-2ece-4221-a278-5d1945d57a57\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.775768 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-catalog-content\") pod \"91661619-2ece-4221-a278-5d1945d57a57\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.775894 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrjb6\" (UniqueName: \"kubernetes.io/projected/91661619-2ece-4221-a278-5d1945d57a57-kube-api-access-hrjb6\") pod \"91661619-2ece-4221-a278-5d1945d57a57\" (UID: \"91661619-2ece-4221-a278-5d1945d57a57\") " Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.776205 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr797\" (UniqueName: \"kubernetes.io/projected/e7631eb4-cbcc-4df8-a4a8-46f00224c41e-kube-api-access-zr797\") on node \"crc\" DevicePath \"\"" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.776407 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-utilities" (OuterVolumeSpecName: "utilities") pod "91661619-2ece-4221-a278-5d1945d57a57" (UID: "91661619-2ece-4221-a278-5d1945d57a57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.784922 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91661619-2ece-4221-a278-5d1945d57a57-kube-api-access-hrjb6" (OuterVolumeSpecName: "kube-api-access-hrjb6") pod "91661619-2ece-4221-a278-5d1945d57a57" (UID: "91661619-2ece-4221-a278-5d1945d57a57"). InnerVolumeSpecName "kube-api-access-hrjb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.877987 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrjb6\" (UniqueName: \"kubernetes.io/projected/91661619-2ece-4221-a278-5d1945d57a57-kube-api-access-hrjb6\") on node \"crc\" DevicePath \"\"" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.878027 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.896294 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91661619-2ece-4221-a278-5d1945d57a57" (UID: "91661619-2ece-4221-a278-5d1945d57a57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:40:04 crc kubenswrapper[4876]: I0313 08:40:04.979780 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91661619-2ece-4221-a278-5d1945d57a57-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.127601 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556520-nm5cz" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.127620 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556520-nm5cz" event={"ID":"e7631eb4-cbcc-4df8-a4a8-46f00224c41e","Type":"ContainerDied","Data":"62f3aafecac63aaa65f8debb53e9542a2771061dc97cf992b4028db056946244"} Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.127707 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62f3aafecac63aaa65f8debb53e9542a2771061dc97cf992b4028db056946244" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.133467 4876 generic.go:334] "Generic (PLEG): container finished" podID="91661619-2ece-4221-a278-5d1945d57a57" containerID="ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01" exitCode=0 Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.133509 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk2h8" event={"ID":"91661619-2ece-4221-a278-5d1945d57a57","Type":"ContainerDied","Data":"ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01"} Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.133541 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk2h8" event={"ID":"91661619-2ece-4221-a278-5d1945d57a57","Type":"ContainerDied","Data":"f14f257f3b26d1f4ae5de9726c1684c7dcaf6403254e3dec0eb65ce19c4162ba"} Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.133541 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rk2h8" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.133578 4876 scope.go:117] "RemoveContainer" containerID="ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.164075 4876 scope.go:117] "RemoveContainer" containerID="d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.168810 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rk2h8"] Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.180567 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rk2h8"] Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.204884 4876 scope.go:117] "RemoveContainer" containerID="f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.244005 4876 scope.go:117] "RemoveContainer" containerID="ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01" Mar 13 08:40:05 crc kubenswrapper[4876]: E0313 08:40:05.244455 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01\": container with ID starting with ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01 not found: ID does not exist" containerID="ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.244495 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01"} err="failed to get container status \"ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01\": rpc error: code = NotFound desc = could not find container \"ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01\": container with ID starting with ca47e5ee9950acc74aa814c9278a5509d1cb07c1a35ed95d507a499fbc3b9c01 not found: ID does not exist" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.244520 4876 scope.go:117] "RemoveContainer" containerID="d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21" Mar 13 08:40:05 crc kubenswrapper[4876]: E0313 08:40:05.244881 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21\": container with ID starting with d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21 not found: ID does not exist" containerID="d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.244923 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21"} err="failed to get container status \"d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21\": rpc error: code = NotFound desc = could not find container \"d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21\": container with ID starting with d5d9b96c5a26e40ae8958b31603d0815cf3f675adcabb9b61c973fc9488afb21 not found: ID does not exist" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.244959 4876 scope.go:117] "RemoveContainer" containerID="f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689" Mar 13 08:40:05 crc kubenswrapper[4876]: E0313 08:40:05.245436 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689\": container with ID starting with f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689 not found: ID does not exist" containerID="f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.245462 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689"} err="failed to get container status \"f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689\": rpc error: code = NotFound desc = could not find container \"f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689\": container with ID starting with f0eaf607fe04f5153ff1ace5ea0c1d39e9124a90299027f813706e9989f73689 not found: ID does not exist" Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.718659 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556514-7mcf6"] Mar 13 08:40:05 crc kubenswrapper[4876]: I0313 08:40:05.733202 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556514-7mcf6"] Mar 13 08:40:07 crc kubenswrapper[4876]: I0313 08:40:07.046522 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d2cee80-a2b8-425d-b58e-54c2710e73a3" path="/var/lib/kubelet/pods/2d2cee80-a2b8-425d-b58e-54c2710e73a3/volumes" Mar 13 08:40:07 crc kubenswrapper[4876]: I0313 08:40:07.048055 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91661619-2ece-4221-a278-5d1945d57a57" path="/var/lib/kubelet/pods/91661619-2ece-4221-a278-5d1945d57a57/volumes" Mar 13 08:40:24 crc kubenswrapper[4876]: I0313 08:40:24.608086 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:40:24 crc kubenswrapper[4876]: I0313 08:40:24.608764 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:40:24 crc kubenswrapper[4876]: I0313 08:40:24.608823 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:40:24 crc kubenswrapper[4876]: I0313 08:40:24.609793 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:40:24 crc kubenswrapper[4876]: I0313 08:40:24.609883 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" gracePeriod=600 Mar 13 08:40:24 crc kubenswrapper[4876]: E0313 08:40:24.733219 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:40:25 crc kubenswrapper[4876]: I0313 08:40:25.314495 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" exitCode=0 Mar 13 08:40:25 crc kubenswrapper[4876]: I0313 08:40:25.314592 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55"} Mar 13 08:40:25 crc kubenswrapper[4876]: I0313 08:40:25.314636 4876 scope.go:117] "RemoveContainer" containerID="cc37175fefd56e14af4bf042273ae967541ee081edd519ea78829a1f09071de2" Mar 13 08:40:25 crc kubenswrapper[4876]: I0313 08:40:25.315505 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:40:25 crc kubenswrapper[4876]: E0313 08:40:25.315890 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.508075 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-22mlj"] Mar 13 08:40:28 crc kubenswrapper[4876]: E0313 08:40:28.509748 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="extract-content" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.509765 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="extract-content" Mar 13 08:40:28 crc kubenswrapper[4876]: E0313 08:40:28.509826 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7631eb4-cbcc-4df8-a4a8-46f00224c41e" containerName="oc" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.509834 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7631eb4-cbcc-4df8-a4a8-46f00224c41e" containerName="oc" Mar 13 08:40:28 crc kubenswrapper[4876]: E0313 08:40:28.509858 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="registry-server" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.509865 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="registry-server" Mar 13 08:40:28 crc kubenswrapper[4876]: E0313 08:40:28.509881 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="extract-utilities" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.509887 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="extract-utilities" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.510064 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="91661619-2ece-4221-a278-5d1945d57a57" containerName="registry-server" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.510084 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7631eb4-cbcc-4df8-a4a8-46f00224c41e" containerName="oc" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.511743 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.521727 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-22mlj"] Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.601579 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-catalog-content\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.601637 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55bwc\" (UniqueName: \"kubernetes.io/projected/d64d62d1-3315-4e9a-8658-59ecedf43f09-kube-api-access-55bwc\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.601960 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-utilities\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.703571 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-catalog-content\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.703664 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55bwc\" (UniqueName: \"kubernetes.io/projected/d64d62d1-3315-4e9a-8658-59ecedf43f09-kube-api-access-55bwc\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.703865 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-utilities\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.704462 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-catalog-content\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.704500 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-utilities\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.728446 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55bwc\" (UniqueName: \"kubernetes.io/projected/d64d62d1-3315-4e9a-8658-59ecedf43f09-kube-api-access-55bwc\") pod \"community-operators-22mlj\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:28 crc kubenswrapper[4876]: I0313 08:40:28.844812 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:29 crc kubenswrapper[4876]: I0313 08:40:29.410326 4876 scope.go:117] "RemoveContainer" containerID="c51464a0d947ef4d66542e316dd62108881d81eb6294e064097a6bcf21099d02" Mar 13 08:40:29 crc kubenswrapper[4876]: I0313 08:40:29.453689 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-22mlj"] Mar 13 08:40:30 crc kubenswrapper[4876]: I0313 08:40:30.374072 4876 generic.go:334] "Generic (PLEG): container finished" podID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerID="a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6" exitCode=0 Mar 13 08:40:30 crc kubenswrapper[4876]: I0313 08:40:30.374210 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22mlj" event={"ID":"d64d62d1-3315-4e9a-8658-59ecedf43f09","Type":"ContainerDied","Data":"a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6"} Mar 13 08:40:30 crc kubenswrapper[4876]: I0313 08:40:30.374573 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22mlj" event={"ID":"d64d62d1-3315-4e9a-8658-59ecedf43f09","Type":"ContainerStarted","Data":"7936ee6db6d2f894b0e32694bd0763f908a2558cec72020e27e213fc28bb141b"} Mar 13 08:40:31 crc kubenswrapper[4876]: I0313 08:40:31.385923 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22mlj" event={"ID":"d64d62d1-3315-4e9a-8658-59ecedf43f09","Type":"ContainerStarted","Data":"b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633"} Mar 13 08:40:32 crc kubenswrapper[4876]: I0313 08:40:32.394781 4876 generic.go:334] "Generic (PLEG): container finished" podID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerID="b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633" exitCode=0 Mar 13 08:40:32 crc kubenswrapper[4876]: I0313 08:40:32.394826 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22mlj" event={"ID":"d64d62d1-3315-4e9a-8658-59ecedf43f09","Type":"ContainerDied","Data":"b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633"} Mar 13 08:40:33 crc kubenswrapper[4876]: I0313 08:40:33.408852 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22mlj" event={"ID":"d64d62d1-3315-4e9a-8658-59ecedf43f09","Type":"ContainerStarted","Data":"2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac"} Mar 13 08:40:33 crc kubenswrapper[4876]: I0313 08:40:33.433574 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-22mlj" podStartSLOduration=2.9549472039999998 podStartE2EDuration="5.433552302s" podCreationTimestamp="2026-03-13 08:40:28 +0000 UTC" firstStartedPulling="2026-03-13 08:40:30.378405002 +0000 UTC m=+3690.049183984" lastFinishedPulling="2026-03-13 08:40:32.85701006 +0000 UTC m=+3692.527789082" observedRunningTime="2026-03-13 08:40:33.429272206 +0000 UTC m=+3693.100051248" watchObservedRunningTime="2026-03-13 08:40:33.433552302 +0000 UTC m=+3693.104331284" Mar 13 08:40:38 crc kubenswrapper[4876]: I0313 08:40:38.036172 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:40:38 crc kubenswrapper[4876]: E0313 08:40:38.036965 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:40:38 crc kubenswrapper[4876]: I0313 08:40:38.845311 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:38 crc kubenswrapper[4876]: I0313 08:40:38.845714 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:38 crc kubenswrapper[4876]: I0313 08:40:38.914129 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:39 crc kubenswrapper[4876]: I0313 08:40:39.513432 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:39 crc kubenswrapper[4876]: I0313 08:40:39.583521 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-22mlj"] Mar 13 08:40:41 crc kubenswrapper[4876]: I0313 08:40:41.495636 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-22mlj" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerName="registry-server" containerID="cri-o://2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac" gracePeriod=2 Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.019543 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.082448 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55bwc\" (UniqueName: \"kubernetes.io/projected/d64d62d1-3315-4e9a-8658-59ecedf43f09-kube-api-access-55bwc\") pod \"d64d62d1-3315-4e9a-8658-59ecedf43f09\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.097216 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64d62d1-3315-4e9a-8658-59ecedf43f09-kube-api-access-55bwc" (OuterVolumeSpecName: "kube-api-access-55bwc") pod "d64d62d1-3315-4e9a-8658-59ecedf43f09" (UID: "d64d62d1-3315-4e9a-8658-59ecedf43f09"). InnerVolumeSpecName "kube-api-access-55bwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.184079 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-catalog-content\") pod \"d64d62d1-3315-4e9a-8658-59ecedf43f09\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.184139 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-utilities\") pod \"d64d62d1-3315-4e9a-8658-59ecedf43f09\" (UID: \"d64d62d1-3315-4e9a-8658-59ecedf43f09\") " Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.184503 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55bwc\" (UniqueName: \"kubernetes.io/projected/d64d62d1-3315-4e9a-8658-59ecedf43f09-kube-api-access-55bwc\") on node \"crc\" DevicePath \"\"" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.185191 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-utilities" (OuterVolumeSpecName: "utilities") pod "d64d62d1-3315-4e9a-8658-59ecedf43f09" (UID: "d64d62d1-3315-4e9a-8658-59ecedf43f09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.239545 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d64d62d1-3315-4e9a-8658-59ecedf43f09" (UID: "d64d62d1-3315-4e9a-8658-59ecedf43f09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.287133 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.287168 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64d62d1-3315-4e9a-8658-59ecedf43f09-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.510855 4876 generic.go:334] "Generic (PLEG): container finished" podID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerID="2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac" exitCode=0 Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.510995 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-22mlj" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.510910 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22mlj" event={"ID":"d64d62d1-3315-4e9a-8658-59ecedf43f09","Type":"ContainerDied","Data":"2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac"} Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.512179 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-22mlj" event={"ID":"d64d62d1-3315-4e9a-8658-59ecedf43f09","Type":"ContainerDied","Data":"7936ee6db6d2f894b0e32694bd0763f908a2558cec72020e27e213fc28bb141b"} Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.512208 4876 scope.go:117] "RemoveContainer" containerID="2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.548932 4876 scope.go:117] "RemoveContainer" containerID="b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.557573 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-22mlj"] Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.569856 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-22mlj"] Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.586063 4876 scope.go:117] "RemoveContainer" containerID="a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.611384 4876 scope.go:117] "RemoveContainer" containerID="2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac" Mar 13 08:40:42 crc kubenswrapper[4876]: E0313 08:40:42.611849 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac\": container with ID starting with 2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac not found: ID does not exist" containerID="2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.611893 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac"} err="failed to get container status \"2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac\": rpc error: code = NotFound desc = could not find container \"2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac\": container with ID starting with 2a5a69ca86ab317475260f3d27f91957f86769bf22eca8b643be3e71645c35ac not found: ID does not exist" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.611921 4876 scope.go:117] "RemoveContainer" containerID="b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633" Mar 13 08:40:42 crc kubenswrapper[4876]: E0313 08:40:42.613569 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633\": container with ID starting with b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633 not found: ID does not exist" containerID="b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.613622 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633"} err="failed to get container status \"b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633\": rpc error: code = NotFound desc = could not find container \"b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633\": container with ID starting with b9bc671e555b44fa298b26ef3892ba4b561d1fcd9b0e3826eacf97c776668633 not found: ID does not exist" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.613652 4876 scope.go:117] "RemoveContainer" containerID="a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6" Mar 13 08:40:42 crc kubenswrapper[4876]: E0313 08:40:42.614130 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6\": container with ID starting with a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6 not found: ID does not exist" containerID="a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6" Mar 13 08:40:42 crc kubenswrapper[4876]: I0313 08:40:42.614163 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6"} err="failed to get container status \"a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6\": rpc error: code = NotFound desc = could not find container \"a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6\": container with ID starting with a5c67318626478190ec7de029ee99b0cfc6ed878a27c7783769482454b42a2a6 not found: ID does not exist" Mar 13 08:40:43 crc kubenswrapper[4876]: I0313 08:40:43.058524 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" path="/var/lib/kubelet/pods/d64d62d1-3315-4e9a-8658-59ecedf43f09/volumes" Mar 13 08:40:50 crc kubenswrapper[4876]: I0313 08:40:50.036159 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:40:50 crc kubenswrapper[4876]: E0313 08:40:50.037093 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:41:01 crc kubenswrapper[4876]: I0313 08:41:01.041946 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:41:01 crc kubenswrapper[4876]: E0313 08:41:01.043055 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:41:13 crc kubenswrapper[4876]: I0313 08:41:13.035783 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:41:13 crc kubenswrapper[4876]: E0313 08:41:13.037261 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:41:24 crc kubenswrapper[4876]: I0313 08:41:24.035622 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:41:24 crc kubenswrapper[4876]: E0313 08:41:24.042450 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:41:35 crc kubenswrapper[4876]: I0313 08:41:35.036078 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:41:35 crc kubenswrapper[4876]: E0313 08:41:35.037761 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:41:37 crc kubenswrapper[4876]: I0313 08:41:37.104957 4876 generic.go:334] "Generic (PLEG): container finished" podID="6ba56888-06b5-406f-b194-0a6e11a72675" containerID="d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069" exitCode=0 Mar 13 08:41:37 crc kubenswrapper[4876]: I0313 08:41:37.105223 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mczk2/must-gather-4mvdh" event={"ID":"6ba56888-06b5-406f-b194-0a6e11a72675","Type":"ContainerDied","Data":"d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069"} Mar 13 08:41:37 crc kubenswrapper[4876]: I0313 08:41:37.105886 4876 scope.go:117] "RemoveContainer" containerID="d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069" Mar 13 08:41:38 crc kubenswrapper[4876]: I0313 08:41:38.029224 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mczk2_must-gather-4mvdh_6ba56888-06b5-406f-b194-0a6e11a72675/gather/0.log" Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.324440 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mczk2/must-gather-4mvdh"] Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.325083 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mczk2/must-gather-4mvdh" podUID="6ba56888-06b5-406f-b194-0a6e11a72675" containerName="copy" containerID="cri-o://72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d" gracePeriod=2 Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.338693 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mczk2/must-gather-4mvdh"] Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.743267 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mczk2_must-gather-4mvdh_6ba56888-06b5-406f-b194-0a6e11a72675/copy/0.log" Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.743872 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.857109 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6ba56888-06b5-406f-b194-0a6e11a72675-must-gather-output\") pod \"6ba56888-06b5-406f-b194-0a6e11a72675\" (UID: \"6ba56888-06b5-406f-b194-0a6e11a72675\") " Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.857261 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7wm7\" (UniqueName: \"kubernetes.io/projected/6ba56888-06b5-406f-b194-0a6e11a72675-kube-api-access-q7wm7\") pod \"6ba56888-06b5-406f-b194-0a6e11a72675\" (UID: \"6ba56888-06b5-406f-b194-0a6e11a72675\") " Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.862496 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba56888-06b5-406f-b194-0a6e11a72675-kube-api-access-q7wm7" (OuterVolumeSpecName: "kube-api-access-q7wm7") pod "6ba56888-06b5-406f-b194-0a6e11a72675" (UID: "6ba56888-06b5-406f-b194-0a6e11a72675"). InnerVolumeSpecName "kube-api-access-q7wm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:41:45 crc kubenswrapper[4876]: I0313 08:41:45.961296 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7wm7\" (UniqueName: \"kubernetes.io/projected/6ba56888-06b5-406f-b194-0a6e11a72675-kube-api-access-q7wm7\") on node \"crc\" DevicePath \"\"" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.010615 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba56888-06b5-406f-b194-0a6e11a72675-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6ba56888-06b5-406f-b194-0a6e11a72675" (UID: "6ba56888-06b5-406f-b194-0a6e11a72675"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.063152 4876 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6ba56888-06b5-406f-b194-0a6e11a72675-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.190091 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mczk2_must-gather-4mvdh_6ba56888-06b5-406f-b194-0a6e11a72675/copy/0.log" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.190834 4876 generic.go:334] "Generic (PLEG): container finished" podID="6ba56888-06b5-406f-b194-0a6e11a72675" containerID="72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d" exitCode=143 Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.190889 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mczk2/must-gather-4mvdh" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.190942 4876 scope.go:117] "RemoveContainer" containerID="72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.212161 4876 scope.go:117] "RemoveContainer" containerID="d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.294477 4876 scope.go:117] "RemoveContainer" containerID="72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d" Mar 13 08:41:46 crc kubenswrapper[4876]: E0313 08:41:46.294874 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d\": container with ID starting with 72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d not found: ID does not exist" containerID="72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.294919 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d"} err="failed to get container status \"72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d\": rpc error: code = NotFound desc = could not find container \"72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d\": container with ID starting with 72e47d6e52f9fb047050b3576264b021f0489f911f95410adc7778aecbc65c7d not found: ID does not exist" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.294948 4876 scope.go:117] "RemoveContainer" containerID="d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069" Mar 13 08:41:46 crc kubenswrapper[4876]: E0313 08:41:46.295443 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069\": container with ID starting with d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069 not found: ID does not exist" containerID="d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069" Mar 13 08:41:46 crc kubenswrapper[4876]: I0313 08:41:46.295538 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069"} err="failed to get container status \"d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069\": rpc error: code = NotFound desc = could not find container \"d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069\": container with ID starting with d7869fe0ac9853ccc1f824dfd658d24d70d3e3952c9be5e25f376e52ff8c7069 not found: ID does not exist" Mar 13 08:41:47 crc kubenswrapper[4876]: I0313 08:41:47.036190 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:41:47 crc kubenswrapper[4876]: E0313 08:41:47.036828 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:41:47 crc kubenswrapper[4876]: I0313 08:41:47.045718 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ba56888-06b5-406f-b194-0a6e11a72675" path="/var/lib/kubelet/pods/6ba56888-06b5-406f-b194-0a6e11a72675/volumes" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.036322 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:42:00 crc kubenswrapper[4876]: E0313 08:42:00.037145 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.153910 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556522-whtwn"] Mar 13 08:42:00 crc kubenswrapper[4876]: E0313 08:42:00.154517 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba56888-06b5-406f-b194-0a6e11a72675" containerName="copy" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.154548 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba56888-06b5-406f-b194-0a6e11a72675" containerName="copy" Mar 13 08:42:00 crc kubenswrapper[4876]: E0313 08:42:00.154587 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba56888-06b5-406f-b194-0a6e11a72675" containerName="gather" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.154600 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba56888-06b5-406f-b194-0a6e11a72675" containerName="gather" Mar 13 08:42:00 crc kubenswrapper[4876]: E0313 08:42:00.154627 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerName="registry-server" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.154641 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerName="registry-server" Mar 13 08:42:00 crc kubenswrapper[4876]: E0313 08:42:00.154670 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerName="extract-utilities" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.154682 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerName="extract-utilities" Mar 13 08:42:00 crc kubenswrapper[4876]: E0313 08:42:00.154708 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerName="extract-content" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.154719 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerName="extract-content" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.155031 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba56888-06b5-406f-b194-0a6e11a72675" containerName="gather" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.155060 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba56888-06b5-406f-b194-0a6e11a72675" containerName="copy" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.155093 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64d62d1-3315-4e9a-8658-59ecedf43f09" containerName="registry-server" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.156096 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556522-whtwn" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.158635 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.158961 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.165105 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556522-whtwn"] Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.167129 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.345288 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smqc2\" (UniqueName: \"kubernetes.io/projected/bb9e0427-5268-4b70-bec0-c1eed76b6554-kube-api-access-smqc2\") pod \"auto-csr-approver-29556522-whtwn\" (UID: \"bb9e0427-5268-4b70-bec0-c1eed76b6554\") " pod="openshift-infra/auto-csr-approver-29556522-whtwn" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.448075 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smqc2\" (UniqueName: \"kubernetes.io/projected/bb9e0427-5268-4b70-bec0-c1eed76b6554-kube-api-access-smqc2\") pod \"auto-csr-approver-29556522-whtwn\" (UID: \"bb9e0427-5268-4b70-bec0-c1eed76b6554\") " pod="openshift-infra/auto-csr-approver-29556522-whtwn" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.466725 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smqc2\" (UniqueName: \"kubernetes.io/projected/bb9e0427-5268-4b70-bec0-c1eed76b6554-kube-api-access-smqc2\") pod \"auto-csr-approver-29556522-whtwn\" (UID: \"bb9e0427-5268-4b70-bec0-c1eed76b6554\") " pod="openshift-infra/auto-csr-approver-29556522-whtwn" Mar 13 08:42:00 crc kubenswrapper[4876]: I0313 08:42:00.485838 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556522-whtwn" Mar 13 08:42:01 crc kubenswrapper[4876]: I0313 08:42:00.999578 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556522-whtwn"] Mar 13 08:42:01 crc kubenswrapper[4876]: I0313 08:42:01.346885 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556522-whtwn" event={"ID":"bb9e0427-5268-4b70-bec0-c1eed76b6554","Type":"ContainerStarted","Data":"e9dce8b842c9143f6785c23779a6e3ccfabd5df775856e5a379bd9cf98f05a19"} Mar 13 08:42:02 crc kubenswrapper[4876]: I0313 08:42:02.358753 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556522-whtwn" event={"ID":"bb9e0427-5268-4b70-bec0-c1eed76b6554","Type":"ContainerStarted","Data":"653e0ba7e8bd41694fd2d300705bc2f149c649dd89711b73d929467fd6d025e6"} Mar 13 08:42:02 crc kubenswrapper[4876]: I0313 08:42:02.379887 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556522-whtwn" podStartSLOduration=1.548467504 podStartE2EDuration="2.379869878s" podCreationTimestamp="2026-03-13 08:42:00 +0000 UTC" firstStartedPulling="2026-03-13 08:42:01.003889528 +0000 UTC m=+3780.674668510" lastFinishedPulling="2026-03-13 08:42:01.835291902 +0000 UTC m=+3781.506070884" observedRunningTime="2026-03-13 08:42:02.37294148 +0000 UTC m=+3782.043720482" watchObservedRunningTime="2026-03-13 08:42:02.379869878 +0000 UTC m=+3782.050648860" Mar 13 08:42:03 crc kubenswrapper[4876]: I0313 08:42:03.370040 4876 generic.go:334] "Generic (PLEG): container finished" podID="bb9e0427-5268-4b70-bec0-c1eed76b6554" containerID="653e0ba7e8bd41694fd2d300705bc2f149c649dd89711b73d929467fd6d025e6" exitCode=0 Mar 13 08:42:03 crc kubenswrapper[4876]: I0313 08:42:03.370105 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556522-whtwn" event={"ID":"bb9e0427-5268-4b70-bec0-c1eed76b6554","Type":"ContainerDied","Data":"653e0ba7e8bd41694fd2d300705bc2f149c649dd89711b73d929467fd6d025e6"} Mar 13 08:42:04 crc kubenswrapper[4876]: I0313 08:42:04.737400 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556522-whtwn" Mar 13 08:42:04 crc kubenswrapper[4876]: I0313 08:42:04.833523 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smqc2\" (UniqueName: \"kubernetes.io/projected/bb9e0427-5268-4b70-bec0-c1eed76b6554-kube-api-access-smqc2\") pod \"bb9e0427-5268-4b70-bec0-c1eed76b6554\" (UID: \"bb9e0427-5268-4b70-bec0-c1eed76b6554\") " Mar 13 08:42:04 crc kubenswrapper[4876]: I0313 08:42:04.840886 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb9e0427-5268-4b70-bec0-c1eed76b6554-kube-api-access-smqc2" (OuterVolumeSpecName: "kube-api-access-smqc2") pod "bb9e0427-5268-4b70-bec0-c1eed76b6554" (UID: "bb9e0427-5268-4b70-bec0-c1eed76b6554"). InnerVolumeSpecName "kube-api-access-smqc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:42:04 crc kubenswrapper[4876]: I0313 08:42:04.935693 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smqc2\" (UniqueName: \"kubernetes.io/projected/bb9e0427-5268-4b70-bec0-c1eed76b6554-kube-api-access-smqc2\") on node \"crc\" DevicePath \"\"" Mar 13 08:42:05 crc kubenswrapper[4876]: I0313 08:42:05.388511 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556522-whtwn" event={"ID":"bb9e0427-5268-4b70-bec0-c1eed76b6554","Type":"ContainerDied","Data":"e9dce8b842c9143f6785c23779a6e3ccfabd5df775856e5a379bd9cf98f05a19"} Mar 13 08:42:05 crc kubenswrapper[4876]: I0313 08:42:05.388560 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9dce8b842c9143f6785c23779a6e3ccfabd5df775856e5a379bd9cf98f05a19" Mar 13 08:42:05 crc kubenswrapper[4876]: I0313 08:42:05.388622 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556522-whtwn" Mar 13 08:42:05 crc kubenswrapper[4876]: I0313 08:42:05.438123 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556516-cqrfr"] Mar 13 08:42:05 crc kubenswrapper[4876]: I0313 08:42:05.447198 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556516-cqrfr"] Mar 13 08:42:07 crc kubenswrapper[4876]: I0313 08:42:07.050904 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5852c991-fc01-46e9-b740-4e716ea321ce" path="/var/lib/kubelet/pods/5852c991-fc01-46e9-b740-4e716ea321ce/volumes" Mar 13 08:42:11 crc kubenswrapper[4876]: I0313 08:42:11.043619 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:42:11 crc kubenswrapper[4876]: E0313 08:42:11.045577 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:42:26 crc kubenswrapper[4876]: I0313 08:42:26.036220 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:42:26 crc kubenswrapper[4876]: E0313 08:42:26.037094 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:42:29 crc kubenswrapper[4876]: I0313 08:42:29.551211 4876 scope.go:117] "RemoveContainer" containerID="cc5382bbe00aef499349eb64c022c168d548f3d969f164f811a26639dc3bc9c8" Mar 13 08:42:29 crc kubenswrapper[4876]: I0313 08:42:29.582707 4876 scope.go:117] "RemoveContainer" containerID="93b509967beab31a37b2e2ceaf3ae117a2d2372bc888e1ed9b4ede0106cbcf82" Mar 13 08:42:37 crc kubenswrapper[4876]: I0313 08:42:37.035282 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:42:37 crc kubenswrapper[4876]: E0313 08:42:37.035913 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:42:48 crc kubenswrapper[4876]: I0313 08:42:48.036365 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:42:48 crc kubenswrapper[4876]: E0313 08:42:48.037700 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:43:00 crc kubenswrapper[4876]: I0313 08:43:00.055623 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:43:00 crc kubenswrapper[4876]: E0313 08:43:00.060287 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:43:11 crc kubenswrapper[4876]: I0313 08:43:11.036542 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:43:11 crc kubenswrapper[4876]: E0313 08:43:11.038320 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:43:25 crc kubenswrapper[4876]: I0313 08:43:25.036364 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:43:25 crc kubenswrapper[4876]: E0313 08:43:25.037218 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:43:40 crc kubenswrapper[4876]: I0313 08:43:40.036579 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:43:40 crc kubenswrapper[4876]: E0313 08:43:40.037991 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:43:51 crc kubenswrapper[4876]: I0313 08:43:51.048855 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:43:51 crc kubenswrapper[4876]: E0313 08:43:51.050073 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.138976 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556524-gqckd"] Mar 13 08:44:00 crc kubenswrapper[4876]: E0313 08:44:00.140096 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb9e0427-5268-4b70-bec0-c1eed76b6554" containerName="oc" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.140110 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb9e0427-5268-4b70-bec0-c1eed76b6554" containerName="oc" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.140327 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb9e0427-5268-4b70-bec0-c1eed76b6554" containerName="oc" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.140961 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556524-gqckd" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.143076 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.143426 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.144113 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.148800 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556524-gqckd"] Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.180655 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d97mg\" (UniqueName: \"kubernetes.io/projected/773a0308-bb40-4075-b62e-2a3abbd04de9-kube-api-access-d97mg\") pod \"auto-csr-approver-29556524-gqckd\" (UID: \"773a0308-bb40-4075-b62e-2a3abbd04de9\") " pod="openshift-infra/auto-csr-approver-29556524-gqckd" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.282562 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d97mg\" (UniqueName: \"kubernetes.io/projected/773a0308-bb40-4075-b62e-2a3abbd04de9-kube-api-access-d97mg\") pod \"auto-csr-approver-29556524-gqckd\" (UID: \"773a0308-bb40-4075-b62e-2a3abbd04de9\") " pod="openshift-infra/auto-csr-approver-29556524-gqckd" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.302474 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d97mg\" (UniqueName: \"kubernetes.io/projected/773a0308-bb40-4075-b62e-2a3abbd04de9-kube-api-access-d97mg\") pod \"auto-csr-approver-29556524-gqckd\" (UID: \"773a0308-bb40-4075-b62e-2a3abbd04de9\") " pod="openshift-infra/auto-csr-approver-29556524-gqckd" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.462951 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556524-gqckd" Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.897381 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556524-gqckd"] Mar 13 08:44:00 crc kubenswrapper[4876]: I0313 08:44:00.899753 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:44:01 crc kubenswrapper[4876]: I0313 08:44:01.579402 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556524-gqckd" event={"ID":"773a0308-bb40-4075-b62e-2a3abbd04de9","Type":"ContainerStarted","Data":"72c5672e797593b6d45f476126e0efc0ae651d0c0604e65b67b25267153b544e"} Mar 13 08:44:02 crc kubenswrapper[4876]: I0313 08:44:02.035754 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:44:02 crc kubenswrapper[4876]: E0313 08:44:02.035998 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:44:02 crc kubenswrapper[4876]: I0313 08:44:02.597485 4876 generic.go:334] "Generic (PLEG): container finished" podID="773a0308-bb40-4075-b62e-2a3abbd04de9" containerID="be7bfa08a56eeacbc1bece6fc207ee653035be7632a9032ec670a6d7a022a0cb" exitCode=0 Mar 13 08:44:02 crc kubenswrapper[4876]: I0313 08:44:02.597782 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556524-gqckd" event={"ID":"773a0308-bb40-4075-b62e-2a3abbd04de9","Type":"ContainerDied","Data":"be7bfa08a56eeacbc1bece6fc207ee653035be7632a9032ec670a6d7a022a0cb"} Mar 13 08:44:03 crc kubenswrapper[4876]: I0313 08:44:03.925295 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556524-gqckd" Mar 13 08:44:03 crc kubenswrapper[4876]: I0313 08:44:03.974466 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d97mg\" (UniqueName: \"kubernetes.io/projected/773a0308-bb40-4075-b62e-2a3abbd04de9-kube-api-access-d97mg\") pod \"773a0308-bb40-4075-b62e-2a3abbd04de9\" (UID: \"773a0308-bb40-4075-b62e-2a3abbd04de9\") " Mar 13 08:44:03 crc kubenswrapper[4876]: I0313 08:44:03.982028 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/773a0308-bb40-4075-b62e-2a3abbd04de9-kube-api-access-d97mg" (OuterVolumeSpecName: "kube-api-access-d97mg") pod "773a0308-bb40-4075-b62e-2a3abbd04de9" (UID: "773a0308-bb40-4075-b62e-2a3abbd04de9"). InnerVolumeSpecName "kube-api-access-d97mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:44:04 crc kubenswrapper[4876]: I0313 08:44:04.076891 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d97mg\" (UniqueName: \"kubernetes.io/projected/773a0308-bb40-4075-b62e-2a3abbd04de9-kube-api-access-d97mg\") on node \"crc\" DevicePath \"\"" Mar 13 08:44:04 crc kubenswrapper[4876]: I0313 08:44:04.619812 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556524-gqckd" event={"ID":"773a0308-bb40-4075-b62e-2a3abbd04de9","Type":"ContainerDied","Data":"72c5672e797593b6d45f476126e0efc0ae651d0c0604e65b67b25267153b544e"} Mar 13 08:44:04 crc kubenswrapper[4876]: I0313 08:44:04.619857 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72c5672e797593b6d45f476126e0efc0ae651d0c0604e65b67b25267153b544e" Mar 13 08:44:04 crc kubenswrapper[4876]: I0313 08:44:04.620082 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556524-gqckd" Mar 13 08:44:05 crc kubenswrapper[4876]: I0313 08:44:05.003353 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556518-5jtkd"] Mar 13 08:44:05 crc kubenswrapper[4876]: I0313 08:44:05.017798 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556518-5jtkd"] Mar 13 08:44:05 crc kubenswrapper[4876]: I0313 08:44:05.047300 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a38b573d-799d-4983-b44b-5f2ec7867aa3" path="/var/lib/kubelet/pods/a38b573d-799d-4983-b44b-5f2ec7867aa3/volumes" Mar 13 08:44:16 crc kubenswrapper[4876]: I0313 08:44:16.035716 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:44:16 crc kubenswrapper[4876]: E0313 08:44:16.036533 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:44:27 crc kubenswrapper[4876]: I0313 08:44:27.036484 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:44:27 crc kubenswrapper[4876]: E0313 08:44:27.037722 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:44:29 crc kubenswrapper[4876]: I0313 08:44:29.733580 4876 scope.go:117] "RemoveContainer" containerID="a15c8d5e1499b9ea2bd1483974ca56f67bbf4fceb6d7b9b9476926d64acae7df" Mar 13 08:44:38 crc kubenswrapper[4876]: I0313 08:44:38.035469 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:44:38 crc kubenswrapper[4876]: E0313 08:44:38.036293 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:44:45 crc kubenswrapper[4876]: I0313 08:44:45.842881 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wgfhk/must-gather-th6r7"] Mar 13 08:44:45 crc kubenswrapper[4876]: E0313 08:44:45.843746 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773a0308-bb40-4075-b62e-2a3abbd04de9" containerName="oc" Mar 13 08:44:45 crc kubenswrapper[4876]: I0313 08:44:45.843757 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="773a0308-bb40-4075-b62e-2a3abbd04de9" containerName="oc" Mar 13 08:44:45 crc kubenswrapper[4876]: I0313 08:44:45.843941 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="773a0308-bb40-4075-b62e-2a3abbd04de9" containerName="oc" Mar 13 08:44:45 crc kubenswrapper[4876]: I0313 08:44:45.865711 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:44:45 crc kubenswrapper[4876]: I0313 08:44:45.867921 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wgfhk/must-gather-th6r7"] Mar 13 08:44:45 crc kubenswrapper[4876]: I0313 08:44:45.868977 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wgfhk"/"default-dockercfg-nx9f6" Mar 13 08:44:45 crc kubenswrapper[4876]: I0313 08:44:45.869349 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wgfhk"/"openshift-service-ca.crt" Mar 13 08:44:45 crc kubenswrapper[4876]: I0313 08:44:45.882958 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wgfhk"/"kube-root-ca.crt" Mar 13 08:44:46 crc kubenswrapper[4876]: I0313 08:44:46.018047 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cc42ec18-ef73-4e58-831e-e6a2ead92f53-must-gather-output\") pod \"must-gather-th6r7\" (UID: \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\") " pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:44:46 crc kubenswrapper[4876]: I0313 08:44:46.018129 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfzz6\" (UniqueName: \"kubernetes.io/projected/cc42ec18-ef73-4e58-831e-e6a2ead92f53-kube-api-access-zfzz6\") pod \"must-gather-th6r7\" (UID: \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\") " pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:44:46 crc kubenswrapper[4876]: I0313 08:44:46.119606 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfzz6\" (UniqueName: \"kubernetes.io/projected/cc42ec18-ef73-4e58-831e-e6a2ead92f53-kube-api-access-zfzz6\") pod \"must-gather-th6r7\" (UID: \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\") " pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:44:46 crc kubenswrapper[4876]: I0313 08:44:46.119846 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cc42ec18-ef73-4e58-831e-e6a2ead92f53-must-gather-output\") pod \"must-gather-th6r7\" (UID: \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\") " pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:44:46 crc kubenswrapper[4876]: I0313 08:44:46.120791 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cc42ec18-ef73-4e58-831e-e6a2ead92f53-must-gather-output\") pod \"must-gather-th6r7\" (UID: \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\") " pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:44:46 crc kubenswrapper[4876]: I0313 08:44:46.139190 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfzz6\" (UniqueName: \"kubernetes.io/projected/cc42ec18-ef73-4e58-831e-e6a2ead92f53-kube-api-access-zfzz6\") pod \"must-gather-th6r7\" (UID: \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\") " pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:44:46 crc kubenswrapper[4876]: I0313 08:44:46.184362 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:44:46 crc kubenswrapper[4876]: I0313 08:44:46.630591 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wgfhk/must-gather-th6r7"] Mar 13 08:44:47 crc kubenswrapper[4876]: I0313 08:44:47.030414 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/must-gather-th6r7" event={"ID":"cc42ec18-ef73-4e58-831e-e6a2ead92f53","Type":"ContainerStarted","Data":"f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35"} Mar 13 08:44:47 crc kubenswrapper[4876]: I0313 08:44:47.030685 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/must-gather-th6r7" event={"ID":"cc42ec18-ef73-4e58-831e-e6a2ead92f53","Type":"ContainerStarted","Data":"8765b954622c053fad8c862f153b00c14e7bb7976e69d99658ca95a7d90ff91b"} Mar 13 08:44:48 crc kubenswrapper[4876]: I0313 08:44:48.047656 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/must-gather-th6r7" event={"ID":"cc42ec18-ef73-4e58-831e-e6a2ead92f53","Type":"ContainerStarted","Data":"eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f"} Mar 13 08:44:49 crc kubenswrapper[4876]: E0313 08:44:49.991297 4876 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.39:60514->38.102.83.39:46711: read tcp 38.102.83.39:60514->38.102.83.39:46711: read: connection reset by peer Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.673997 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wgfhk/must-gather-th6r7" podStartSLOduration=5.673975484 podStartE2EDuration="5.673975484s" podCreationTimestamp="2026-03-13 08:44:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:44:48.084453645 +0000 UTC m=+3947.755232637" watchObservedRunningTime="2026-03-13 08:44:50.673975484 +0000 UTC m=+3950.344754466" Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.678846 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-qp7ht"] Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.680054 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.818228 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmpzz\" (UniqueName: \"kubernetes.io/projected/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-kube-api-access-tmpzz\") pod \"crc-debug-qp7ht\" (UID: \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\") " pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.818310 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-host\") pod \"crc-debug-qp7ht\" (UID: \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\") " pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.919757 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmpzz\" (UniqueName: \"kubernetes.io/projected/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-kube-api-access-tmpzz\") pod \"crc-debug-qp7ht\" (UID: \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\") " pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.919812 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-host\") pod \"crc-debug-qp7ht\" (UID: \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\") " pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.920038 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-host\") pod \"crc-debug-qp7ht\" (UID: \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\") " pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.938081 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmpzz\" (UniqueName: \"kubernetes.io/projected/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-kube-api-access-tmpzz\") pod \"crc-debug-qp7ht\" (UID: \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\") " pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:44:50 crc kubenswrapper[4876]: I0313 08:44:50.998757 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:44:51 crc kubenswrapper[4876]: W0313 08:44:51.033292 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fa743ce_0fc8_40bf_b79e_0e34b37f0006.slice/crio-e5bb05cddaed1d723e0e5b959ecc85e47f2af90e1b9d84da6e952e8d8a0b40fe WatchSource:0}: Error finding container e5bb05cddaed1d723e0e5b959ecc85e47f2af90e1b9d84da6e952e8d8a0b40fe: Status 404 returned error can't find the container with id e5bb05cddaed1d723e0e5b959ecc85e47f2af90e1b9d84da6e952e8d8a0b40fe Mar 13 08:44:51 crc kubenswrapper[4876]: I0313 08:44:51.053574 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:44:51 crc kubenswrapper[4876]: E0313 08:44:51.054713 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:44:51 crc kubenswrapper[4876]: I0313 08:44:51.081112 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" event={"ID":"2fa743ce-0fc8-40bf-b79e-0e34b37f0006","Type":"ContainerStarted","Data":"e5bb05cddaed1d723e0e5b959ecc85e47f2af90e1b9d84da6e952e8d8a0b40fe"} Mar 13 08:44:52 crc kubenswrapper[4876]: I0313 08:44:52.093716 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" event={"ID":"2fa743ce-0fc8-40bf-b79e-0e34b37f0006","Type":"ContainerStarted","Data":"b4af3c335586b0e7ef8f2e5ec1057693f9c65dd3a951a20c481f5d300c8524ac"} Mar 13 08:44:52 crc kubenswrapper[4876]: I0313 08:44:52.109480 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" podStartSLOduration=2.10944955 podStartE2EDuration="2.10944955s" podCreationTimestamp="2026-03-13 08:44:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 08:44:52.10616147 +0000 UTC m=+3951.776940452" watchObservedRunningTime="2026-03-13 08:44:52.10944955 +0000 UTC m=+3951.780228572" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.158885 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d"] Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.161162 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.164708 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.164808 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.174564 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d"] Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.333988 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/840a79dc-25bc-4d60-9945-34e013433c59-secret-volume\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.334099 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2btf\" (UniqueName: \"kubernetes.io/projected/840a79dc-25bc-4d60-9945-34e013433c59-kube-api-access-v2btf\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.334749 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/840a79dc-25bc-4d60-9945-34e013433c59-config-volume\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.437264 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/840a79dc-25bc-4d60-9945-34e013433c59-config-volume\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.437405 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/840a79dc-25bc-4d60-9945-34e013433c59-secret-volume\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.437471 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2btf\" (UniqueName: \"kubernetes.io/projected/840a79dc-25bc-4d60-9945-34e013433c59-kube-api-access-v2btf\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.438618 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/840a79dc-25bc-4d60-9945-34e013433c59-config-volume\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.458898 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/840a79dc-25bc-4d60-9945-34e013433c59-secret-volume\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.462722 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2btf\" (UniqueName: \"kubernetes.io/projected/840a79dc-25bc-4d60-9945-34e013433c59-kube-api-access-v2btf\") pod \"collect-profiles-29556525-sgr7d\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:00 crc kubenswrapper[4876]: I0313 08:45:00.553361 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:01 crc kubenswrapper[4876]: I0313 08:45:01.081334 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d"] Mar 13 08:45:01 crc kubenswrapper[4876]: I0313 08:45:01.211400 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" event={"ID":"840a79dc-25bc-4d60-9945-34e013433c59","Type":"ContainerStarted","Data":"e6cdd983bc477e5e290314c6c8fa5e2a836e677c8d2b192d1c435fa7e60c07a9"} Mar 13 08:45:02 crc kubenswrapper[4876]: I0313 08:45:02.220955 4876 generic.go:334] "Generic (PLEG): container finished" podID="840a79dc-25bc-4d60-9945-34e013433c59" containerID="51ee1cc91f837827ec05ac1b17bdc6f4868bb7dd801bae44558a50f7b50154dd" exitCode=0 Mar 13 08:45:02 crc kubenswrapper[4876]: I0313 08:45:02.221184 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" event={"ID":"840a79dc-25bc-4d60-9945-34e013433c59","Type":"ContainerDied","Data":"51ee1cc91f837827ec05ac1b17bdc6f4868bb7dd801bae44558a50f7b50154dd"} Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.041759 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:45:03 crc kubenswrapper[4876]: E0313 08:45:03.047441 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.626782 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.804892 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/840a79dc-25bc-4d60-9945-34e013433c59-config-volume\") pod \"840a79dc-25bc-4d60-9945-34e013433c59\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.804980 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/840a79dc-25bc-4d60-9945-34e013433c59-secret-volume\") pod \"840a79dc-25bc-4d60-9945-34e013433c59\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.805052 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2btf\" (UniqueName: \"kubernetes.io/projected/840a79dc-25bc-4d60-9945-34e013433c59-kube-api-access-v2btf\") pod \"840a79dc-25bc-4d60-9945-34e013433c59\" (UID: \"840a79dc-25bc-4d60-9945-34e013433c59\") " Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.805728 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840a79dc-25bc-4d60-9945-34e013433c59-config-volume" (OuterVolumeSpecName: "config-volume") pod "840a79dc-25bc-4d60-9945-34e013433c59" (UID: "840a79dc-25bc-4d60-9945-34e013433c59"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.815413 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840a79dc-25bc-4d60-9945-34e013433c59-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "840a79dc-25bc-4d60-9945-34e013433c59" (UID: "840a79dc-25bc-4d60-9945-34e013433c59"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.815507 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/840a79dc-25bc-4d60-9945-34e013433c59-kube-api-access-v2btf" (OuterVolumeSpecName: "kube-api-access-v2btf") pod "840a79dc-25bc-4d60-9945-34e013433c59" (UID: "840a79dc-25bc-4d60-9945-34e013433c59"). InnerVolumeSpecName "kube-api-access-v2btf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.907729 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2btf\" (UniqueName: \"kubernetes.io/projected/840a79dc-25bc-4d60-9945-34e013433c59-kube-api-access-v2btf\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.908193 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/840a79dc-25bc-4d60-9945-34e013433c59-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:03 crc kubenswrapper[4876]: I0313 08:45:03.908285 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/840a79dc-25bc-4d60-9945-34e013433c59-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:04 crc kubenswrapper[4876]: I0313 08:45:04.239046 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" event={"ID":"840a79dc-25bc-4d60-9945-34e013433c59","Type":"ContainerDied","Data":"e6cdd983bc477e5e290314c6c8fa5e2a836e677c8d2b192d1c435fa7e60c07a9"} Mar 13 08:45:04 crc kubenswrapper[4876]: I0313 08:45:04.239359 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6cdd983bc477e5e290314c6c8fa5e2a836e677c8d2b192d1c435fa7e60c07a9" Mar 13 08:45:04 crc kubenswrapper[4876]: I0313 08:45:04.239427 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556525-sgr7d" Mar 13 08:45:04 crc kubenswrapper[4876]: I0313 08:45:04.709545 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd"] Mar 13 08:45:04 crc kubenswrapper[4876]: I0313 08:45:04.718003 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556480-q2pqd"] Mar 13 08:45:05 crc kubenswrapper[4876]: I0313 08:45:05.054112 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="556fa6a5-90ac-43ea-a365-abd7ea03f21a" path="/var/lib/kubelet/pods/556fa6a5-90ac-43ea-a365-abd7ea03f21a/volumes" Mar 13 08:45:16 crc kubenswrapper[4876]: I0313 08:45:16.035947 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:45:16 crc kubenswrapper[4876]: E0313 08:45:16.036673 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:45:27 crc kubenswrapper[4876]: I0313 08:45:27.035409 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:45:27 crc kubenswrapper[4876]: I0313 08:45:27.433875 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"5f102aa15a16bff8287b68350bdd63c56ec0a2889615f66559f3f6993ac5c24a"} Mar 13 08:45:27 crc kubenswrapper[4876]: I0313 08:45:27.436016 4876 generic.go:334] "Generic (PLEG): container finished" podID="2fa743ce-0fc8-40bf-b79e-0e34b37f0006" containerID="b4af3c335586b0e7ef8f2e5ec1057693f9c65dd3a951a20c481f5d300c8524ac" exitCode=0 Mar 13 08:45:27 crc kubenswrapper[4876]: I0313 08:45:27.436084 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" event={"ID":"2fa743ce-0fc8-40bf-b79e-0e34b37f0006","Type":"ContainerDied","Data":"b4af3c335586b0e7ef8f2e5ec1057693f9c65dd3a951a20c481f5d300c8524ac"} Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.547001 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.587656 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-qp7ht"] Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.600698 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-qp7ht"] Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.669203 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-host\") pod \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\" (UID: \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\") " Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.669461 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmpzz\" (UniqueName: \"kubernetes.io/projected/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-kube-api-access-tmpzz\") pod \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\" (UID: \"2fa743ce-0fc8-40bf-b79e-0e34b37f0006\") " Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.669585 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-host" (OuterVolumeSpecName: "host") pod "2fa743ce-0fc8-40bf-b79e-0e34b37f0006" (UID: "2fa743ce-0fc8-40bf-b79e-0e34b37f0006"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.669971 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-host\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.715430 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-kube-api-access-tmpzz" (OuterVolumeSpecName: "kube-api-access-tmpzz") pod "2fa743ce-0fc8-40bf-b79e-0e34b37f0006" (UID: "2fa743ce-0fc8-40bf-b79e-0e34b37f0006"). InnerVolumeSpecName "kube-api-access-tmpzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.749259 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8g5x8"] Mar 13 08:45:28 crc kubenswrapper[4876]: E0313 08:45:28.749674 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa743ce-0fc8-40bf-b79e-0e34b37f0006" containerName="container-00" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.749698 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa743ce-0fc8-40bf-b79e-0e34b37f0006" containerName="container-00" Mar 13 08:45:28 crc kubenswrapper[4876]: E0313 08:45:28.749734 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840a79dc-25bc-4d60-9945-34e013433c59" containerName="collect-profiles" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.749740 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="840a79dc-25bc-4d60-9945-34e013433c59" containerName="collect-profiles" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.749940 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="840a79dc-25bc-4d60-9945-34e013433c59" containerName="collect-profiles" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.749977 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fa743ce-0fc8-40bf-b79e-0e34b37f0006" containerName="container-00" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.751369 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.764860 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8g5x8"] Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.771498 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmpzz\" (UniqueName: \"kubernetes.io/projected/2fa743ce-0fc8-40bf-b79e-0e34b37f0006-kube-api-access-tmpzz\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.873021 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-utilities\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.873539 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-catalog-content\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.873857 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv7qv\" (UniqueName: \"kubernetes.io/projected/44011e32-c93d-44d2-93ae-5d8a3be8761e-kube-api-access-lv7qv\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.975230 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv7qv\" (UniqueName: \"kubernetes.io/projected/44011e32-c93d-44d2-93ae-5d8a3be8761e-kube-api-access-lv7qv\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.975329 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-utilities\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.975405 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-catalog-content\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.975875 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-utilities\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:28 crc kubenswrapper[4876]: I0313 08:45:28.975883 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-catalog-content\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:29 crc kubenswrapper[4876]: I0313 08:45:28.995055 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv7qv\" (UniqueName: \"kubernetes.io/projected/44011e32-c93d-44d2-93ae-5d8a3be8761e-kube-api-access-lv7qv\") pod \"certified-operators-8g5x8\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:29 crc kubenswrapper[4876]: I0313 08:45:29.048583 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fa743ce-0fc8-40bf-b79e-0e34b37f0006" path="/var/lib/kubelet/pods/2fa743ce-0fc8-40bf-b79e-0e34b37f0006/volumes" Mar 13 08:45:29 crc kubenswrapper[4876]: I0313 08:45:29.091203 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:29 crc kubenswrapper[4876]: I0313 08:45:29.461357 4876 scope.go:117] "RemoveContainer" containerID="b4af3c335586b0e7ef8f2e5ec1057693f9c65dd3a951a20c481f5d300c8524ac" Mar 13 08:45:29 crc kubenswrapper[4876]: I0313 08:45:29.461375 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-qp7ht" Mar 13 08:45:29 crc kubenswrapper[4876]: I0313 08:45:29.654684 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8g5x8"] Mar 13 08:45:29 crc kubenswrapper[4876]: I0313 08:45:29.824339 4876 scope.go:117] "RemoveContainer" containerID="f98f31af3ff8ab23b2d9f95916001dd27779a009e580066cf5aafb24a665bc37" Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.014082 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-jv2qv"] Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.015499 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.095244 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86a2aa33-67d4-43d2-a308-f65fff16fc62-host\") pod \"crc-debug-jv2qv\" (UID: \"86a2aa33-67d4-43d2-a308-f65fff16fc62\") " pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.095441 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49pn6\" (UniqueName: \"kubernetes.io/projected/86a2aa33-67d4-43d2-a308-f65fff16fc62-kube-api-access-49pn6\") pod \"crc-debug-jv2qv\" (UID: \"86a2aa33-67d4-43d2-a308-f65fff16fc62\") " pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.196702 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49pn6\" (UniqueName: \"kubernetes.io/projected/86a2aa33-67d4-43d2-a308-f65fff16fc62-kube-api-access-49pn6\") pod \"crc-debug-jv2qv\" (UID: \"86a2aa33-67d4-43d2-a308-f65fff16fc62\") " pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.196825 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86a2aa33-67d4-43d2-a308-f65fff16fc62-host\") pod \"crc-debug-jv2qv\" (UID: \"86a2aa33-67d4-43d2-a308-f65fff16fc62\") " pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.196979 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86a2aa33-67d4-43d2-a308-f65fff16fc62-host\") pod \"crc-debug-jv2qv\" (UID: \"86a2aa33-67d4-43d2-a308-f65fff16fc62\") " pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.214403 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49pn6\" (UniqueName: \"kubernetes.io/projected/86a2aa33-67d4-43d2-a308-f65fff16fc62-kube-api-access-49pn6\") pod \"crc-debug-jv2qv\" (UID: \"86a2aa33-67d4-43d2-a308-f65fff16fc62\") " pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.330094 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:30 crc kubenswrapper[4876]: W0313 08:45:30.355009 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86a2aa33_67d4_43d2_a308_f65fff16fc62.slice/crio-01c4d5bac11ea29a7cfb2f18b440dd282ed0ab09881f57e15802ecd5634b8a86 WatchSource:0}: Error finding container 01c4d5bac11ea29a7cfb2f18b440dd282ed0ab09881f57e15802ecd5634b8a86: Status 404 returned error can't find the container with id 01c4d5bac11ea29a7cfb2f18b440dd282ed0ab09881f57e15802ecd5634b8a86 Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.474530 4876 generic.go:334] "Generic (PLEG): container finished" podID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerID="79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320" exitCode=0 Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.474913 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g5x8" event={"ID":"44011e32-c93d-44d2-93ae-5d8a3be8761e","Type":"ContainerDied","Data":"79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320"} Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.474947 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g5x8" event={"ID":"44011e32-c93d-44d2-93ae-5d8a3be8761e","Type":"ContainerStarted","Data":"f6e6bdaf5a0b82aa069907e1a98dfcfd0664d6f54b9c467088caaa4826bca546"} Mar 13 08:45:30 crc kubenswrapper[4876]: I0313 08:45:30.477676 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" event={"ID":"86a2aa33-67d4-43d2-a308-f65fff16fc62","Type":"ContainerStarted","Data":"01c4d5bac11ea29a7cfb2f18b440dd282ed0ab09881f57e15802ecd5634b8a86"} Mar 13 08:45:31 crc kubenswrapper[4876]: I0313 08:45:31.487532 4876 generic.go:334] "Generic (PLEG): container finished" podID="86a2aa33-67d4-43d2-a308-f65fff16fc62" containerID="e7de68df3afd44e8a711b70f00754a82532b97ef166d0f267c6bcd64e3f5bb8d" exitCode=0 Mar 13 08:45:31 crc kubenswrapper[4876]: I0313 08:45:31.487580 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" event={"ID":"86a2aa33-67d4-43d2-a308-f65fff16fc62","Type":"ContainerDied","Data":"e7de68df3afd44e8a711b70f00754a82532b97ef166d0f267c6bcd64e3f5bb8d"} Mar 13 08:45:31 crc kubenswrapper[4876]: I0313 08:45:31.861541 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-jv2qv"] Mar 13 08:45:31 crc kubenswrapper[4876]: I0313 08:45:31.877586 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-jv2qv"] Mar 13 08:45:32 crc kubenswrapper[4876]: I0313 08:45:32.497916 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g5x8" event={"ID":"44011e32-c93d-44d2-93ae-5d8a3be8761e","Type":"ContainerStarted","Data":"d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d"} Mar 13 08:45:32 crc kubenswrapper[4876]: I0313 08:45:32.594819 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:32 crc kubenswrapper[4876]: I0313 08:45:32.648218 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49pn6\" (UniqueName: \"kubernetes.io/projected/86a2aa33-67d4-43d2-a308-f65fff16fc62-kube-api-access-49pn6\") pod \"86a2aa33-67d4-43d2-a308-f65fff16fc62\" (UID: \"86a2aa33-67d4-43d2-a308-f65fff16fc62\") " Mar 13 08:45:32 crc kubenswrapper[4876]: I0313 08:45:32.648398 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86a2aa33-67d4-43d2-a308-f65fff16fc62-host\") pod \"86a2aa33-67d4-43d2-a308-f65fff16fc62\" (UID: \"86a2aa33-67d4-43d2-a308-f65fff16fc62\") " Mar 13 08:45:32 crc kubenswrapper[4876]: I0313 08:45:32.648510 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86a2aa33-67d4-43d2-a308-f65fff16fc62-host" (OuterVolumeSpecName: "host") pod "86a2aa33-67d4-43d2-a308-f65fff16fc62" (UID: "86a2aa33-67d4-43d2-a308-f65fff16fc62"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 08:45:32 crc kubenswrapper[4876]: I0313 08:45:32.648905 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86a2aa33-67d4-43d2-a308-f65fff16fc62-host\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:32 crc kubenswrapper[4876]: I0313 08:45:32.654358 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86a2aa33-67d4-43d2-a308-f65fff16fc62-kube-api-access-49pn6" (OuterVolumeSpecName: "kube-api-access-49pn6") pod "86a2aa33-67d4-43d2-a308-f65fff16fc62" (UID: "86a2aa33-67d4-43d2-a308-f65fff16fc62"). InnerVolumeSpecName "kube-api-access-49pn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:45:32 crc kubenswrapper[4876]: I0313 08:45:32.751205 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49pn6\" (UniqueName: \"kubernetes.io/projected/86a2aa33-67d4-43d2-a308-f65fff16fc62-kube-api-access-49pn6\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.052975 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86a2aa33-67d4-43d2-a308-f65fff16fc62" path="/var/lib/kubelet/pods/86a2aa33-67d4-43d2-a308-f65fff16fc62/volumes" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.345720 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-g4h2h"] Mar 13 08:45:33 crc kubenswrapper[4876]: E0313 08:45:33.346102 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a2aa33-67d4-43d2-a308-f65fff16fc62" containerName="container-00" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.346119 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a2aa33-67d4-43d2-a308-f65fff16fc62" containerName="container-00" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.346371 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="86a2aa33-67d4-43d2-a308-f65fff16fc62" containerName="container-00" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.347070 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.462162 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/673c9f92-1c36-445c-8566-aba08dc8a304-host\") pod \"crc-debug-g4h2h\" (UID: \"673c9f92-1c36-445c-8566-aba08dc8a304\") " pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.462435 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd6q2\" (UniqueName: \"kubernetes.io/projected/673c9f92-1c36-445c-8566-aba08dc8a304-kube-api-access-vd6q2\") pod \"crc-debug-g4h2h\" (UID: \"673c9f92-1c36-445c-8566-aba08dc8a304\") " pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.527004 4876 generic.go:334] "Generic (PLEG): container finished" podID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerID="d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d" exitCode=0 Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.527125 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g5x8" event={"ID":"44011e32-c93d-44d2-93ae-5d8a3be8761e","Type":"ContainerDied","Data":"d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d"} Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.529118 4876 scope.go:117] "RemoveContainer" containerID="e7de68df3afd44e8a711b70f00754a82532b97ef166d0f267c6bcd64e3f5bb8d" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.529159 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-jv2qv" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.564413 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/673c9f92-1c36-445c-8566-aba08dc8a304-host\") pod \"crc-debug-g4h2h\" (UID: \"673c9f92-1c36-445c-8566-aba08dc8a304\") " pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.564504 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd6q2\" (UniqueName: \"kubernetes.io/projected/673c9f92-1c36-445c-8566-aba08dc8a304-kube-api-access-vd6q2\") pod \"crc-debug-g4h2h\" (UID: \"673c9f92-1c36-445c-8566-aba08dc8a304\") " pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.565573 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/673c9f92-1c36-445c-8566-aba08dc8a304-host\") pod \"crc-debug-g4h2h\" (UID: \"673c9f92-1c36-445c-8566-aba08dc8a304\") " pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.603384 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd6q2\" (UniqueName: \"kubernetes.io/projected/673c9f92-1c36-445c-8566-aba08dc8a304-kube-api-access-vd6q2\") pod \"crc-debug-g4h2h\" (UID: \"673c9f92-1c36-445c-8566-aba08dc8a304\") " pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:33 crc kubenswrapper[4876]: I0313 08:45:33.665027 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:33 crc kubenswrapper[4876]: W0313 08:45:33.693574 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod673c9f92_1c36_445c_8566_aba08dc8a304.slice/crio-7f0a0f3c7ab1e12d76db44878aba4a1f01ee74d53762fc2f586f61171de21732 WatchSource:0}: Error finding container 7f0a0f3c7ab1e12d76db44878aba4a1f01ee74d53762fc2f586f61171de21732: Status 404 returned error can't find the container with id 7f0a0f3c7ab1e12d76db44878aba4a1f01ee74d53762fc2f586f61171de21732 Mar 13 08:45:34 crc kubenswrapper[4876]: I0313 08:45:34.538905 4876 generic.go:334] "Generic (PLEG): container finished" podID="673c9f92-1c36-445c-8566-aba08dc8a304" containerID="6b07724b127e17104892c2bc33e1c442c4a08693ecdcd550b4f6e3c38551f9e4" exitCode=0 Mar 13 08:45:34 crc kubenswrapper[4876]: I0313 08:45:34.539007 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" event={"ID":"673c9f92-1c36-445c-8566-aba08dc8a304","Type":"ContainerDied","Data":"6b07724b127e17104892c2bc33e1c442c4a08693ecdcd550b4f6e3c38551f9e4"} Mar 13 08:45:34 crc kubenswrapper[4876]: I0313 08:45:34.539380 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" event={"ID":"673c9f92-1c36-445c-8566-aba08dc8a304","Type":"ContainerStarted","Data":"7f0a0f3c7ab1e12d76db44878aba4a1f01ee74d53762fc2f586f61171de21732"} Mar 13 08:45:34 crc kubenswrapper[4876]: I0313 08:45:34.544303 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g5x8" event={"ID":"44011e32-c93d-44d2-93ae-5d8a3be8761e","Type":"ContainerStarted","Data":"a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4"} Mar 13 08:45:34 crc kubenswrapper[4876]: I0313 08:45:34.575124 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8g5x8" podStartSLOduration=2.8818520789999997 podStartE2EDuration="6.575101461s" podCreationTimestamp="2026-03-13 08:45:28 +0000 UTC" firstStartedPulling="2026-03-13 08:45:30.476551225 +0000 UTC m=+3990.147330207" lastFinishedPulling="2026-03-13 08:45:34.169800607 +0000 UTC m=+3993.840579589" observedRunningTime="2026-03-13 08:45:34.570565987 +0000 UTC m=+3994.241344969" watchObservedRunningTime="2026-03-13 08:45:34.575101461 +0000 UTC m=+3994.245880443" Mar 13 08:45:34 crc kubenswrapper[4876]: I0313 08:45:34.596443 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-g4h2h"] Mar 13 08:45:34 crc kubenswrapper[4876]: I0313 08:45:34.605150 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wgfhk/crc-debug-g4h2h"] Mar 13 08:45:35 crc kubenswrapper[4876]: I0313 08:45:35.668136 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:35 crc kubenswrapper[4876]: I0313 08:45:35.702836 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/673c9f92-1c36-445c-8566-aba08dc8a304-host\") pod \"673c9f92-1c36-445c-8566-aba08dc8a304\" (UID: \"673c9f92-1c36-445c-8566-aba08dc8a304\") " Mar 13 08:45:35 crc kubenswrapper[4876]: I0313 08:45:35.703047 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd6q2\" (UniqueName: \"kubernetes.io/projected/673c9f92-1c36-445c-8566-aba08dc8a304-kube-api-access-vd6q2\") pod \"673c9f92-1c36-445c-8566-aba08dc8a304\" (UID: \"673c9f92-1c36-445c-8566-aba08dc8a304\") " Mar 13 08:45:35 crc kubenswrapper[4876]: I0313 08:45:35.703161 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/673c9f92-1c36-445c-8566-aba08dc8a304-host" (OuterVolumeSpecName: "host") pod "673c9f92-1c36-445c-8566-aba08dc8a304" (UID: "673c9f92-1c36-445c-8566-aba08dc8a304"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 08:45:35 crc kubenswrapper[4876]: I0313 08:45:35.703535 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/673c9f92-1c36-445c-8566-aba08dc8a304-host\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:35 crc kubenswrapper[4876]: I0313 08:45:35.711367 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673c9f92-1c36-445c-8566-aba08dc8a304-kube-api-access-vd6q2" (OuterVolumeSpecName: "kube-api-access-vd6q2") pod "673c9f92-1c36-445c-8566-aba08dc8a304" (UID: "673c9f92-1c36-445c-8566-aba08dc8a304"). InnerVolumeSpecName "kube-api-access-vd6q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:45:35 crc kubenswrapper[4876]: I0313 08:45:35.805479 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd6q2\" (UniqueName: \"kubernetes.io/projected/673c9f92-1c36-445c-8566-aba08dc8a304-kube-api-access-vd6q2\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:36 crc kubenswrapper[4876]: I0313 08:45:36.563486 4876 scope.go:117] "RemoveContainer" containerID="6b07724b127e17104892c2bc33e1c442c4a08693ecdcd550b4f6e3c38551f9e4" Mar 13 08:45:36 crc kubenswrapper[4876]: I0313 08:45:36.563538 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/crc-debug-g4h2h" Mar 13 08:45:37 crc kubenswrapper[4876]: I0313 08:45:37.048115 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="673c9f92-1c36-445c-8566-aba08dc8a304" path="/var/lib/kubelet/pods/673c9f92-1c36-445c-8566-aba08dc8a304/volumes" Mar 13 08:45:39 crc kubenswrapper[4876]: I0313 08:45:39.091931 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:39 crc kubenswrapper[4876]: I0313 08:45:39.094114 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:39 crc kubenswrapper[4876]: I0313 08:45:39.155927 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:39 crc kubenswrapper[4876]: I0313 08:45:39.642901 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:40 crc kubenswrapper[4876]: I0313 08:45:40.952744 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8g5x8"] Mar 13 08:45:41 crc kubenswrapper[4876]: I0313 08:45:41.613841 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8g5x8" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerName="registry-server" containerID="cri-o://a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4" gracePeriod=2 Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.174751 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.229013 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-catalog-content\") pod \"44011e32-c93d-44d2-93ae-5d8a3be8761e\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.229409 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv7qv\" (UniqueName: \"kubernetes.io/projected/44011e32-c93d-44d2-93ae-5d8a3be8761e-kube-api-access-lv7qv\") pod \"44011e32-c93d-44d2-93ae-5d8a3be8761e\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.229581 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-utilities\") pod \"44011e32-c93d-44d2-93ae-5d8a3be8761e\" (UID: \"44011e32-c93d-44d2-93ae-5d8a3be8761e\") " Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.231032 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-utilities" (OuterVolumeSpecName: "utilities") pod "44011e32-c93d-44d2-93ae-5d8a3be8761e" (UID: "44011e32-c93d-44d2-93ae-5d8a3be8761e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.273144 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44011e32-c93d-44d2-93ae-5d8a3be8761e-kube-api-access-lv7qv" (OuterVolumeSpecName: "kube-api-access-lv7qv") pod "44011e32-c93d-44d2-93ae-5d8a3be8761e" (UID: "44011e32-c93d-44d2-93ae-5d8a3be8761e"). InnerVolumeSpecName "kube-api-access-lv7qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.332446 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.332480 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv7qv\" (UniqueName: \"kubernetes.io/projected/44011e32-c93d-44d2-93ae-5d8a3be8761e-kube-api-access-lv7qv\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.624948 4876 generic.go:334] "Generic (PLEG): container finished" podID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerID="a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4" exitCode=0 Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.624994 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g5x8" event={"ID":"44011e32-c93d-44d2-93ae-5d8a3be8761e","Type":"ContainerDied","Data":"a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4"} Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.625021 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g5x8" event={"ID":"44011e32-c93d-44d2-93ae-5d8a3be8761e","Type":"ContainerDied","Data":"f6e6bdaf5a0b82aa069907e1a98dfcfd0664d6f54b9c467088caaa4826bca546"} Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.625038 4876 scope.go:117] "RemoveContainer" containerID="a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.625062 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8g5x8" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.646482 4876 scope.go:117] "RemoveContainer" containerID="d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.667094 4876 scope.go:117] "RemoveContainer" containerID="79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.701339 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44011e32-c93d-44d2-93ae-5d8a3be8761e" (UID: "44011e32-c93d-44d2-93ae-5d8a3be8761e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.725509 4876 scope.go:117] "RemoveContainer" containerID="a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4" Mar 13 08:45:42 crc kubenswrapper[4876]: E0313 08:45:42.725978 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4\": container with ID starting with a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4 not found: ID does not exist" containerID="a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.726027 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4"} err="failed to get container status \"a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4\": rpc error: code = NotFound desc = could not find container \"a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4\": container with ID starting with a12ad14deabec19e88b9b800f922e73e2c9d623a3308b7067ac4629d9aa49de4 not found: ID does not exist" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.726108 4876 scope.go:117] "RemoveContainer" containerID="d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d" Mar 13 08:45:42 crc kubenswrapper[4876]: E0313 08:45:42.726437 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d\": container with ID starting with d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d not found: ID does not exist" containerID="d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.726480 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d"} err="failed to get container status \"d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d\": rpc error: code = NotFound desc = could not find container \"d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d\": container with ID starting with d0bd3c59c0110c0602f3b354ba82c916e31c51dfe21ecd224ed84b3659ca274d not found: ID does not exist" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.726507 4876 scope.go:117] "RemoveContainer" containerID="79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320" Mar 13 08:45:42 crc kubenswrapper[4876]: E0313 08:45:42.729229 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320\": container with ID starting with 79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320 not found: ID does not exist" containerID="79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.729276 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320"} err="failed to get container status \"79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320\": rpc error: code = NotFound desc = could not find container \"79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320\": container with ID starting with 79bbde4f9a2440cb79e0f0b080643a47fa9a9188af3cce716cf5602063e7e320 not found: ID does not exist" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.741505 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44011e32-c93d-44d2-93ae-5d8a3be8761e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.972977 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8g5x8"] Mar 13 08:45:42 crc kubenswrapper[4876]: I0313 08:45:42.985422 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8g5x8"] Mar 13 08:45:43 crc kubenswrapper[4876]: I0313 08:45:43.049906 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" path="/var/lib/kubelet/pods/44011e32-c93d-44d2-93ae-5d8a3be8761e/volumes" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.148645 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556526-cgtsw"] Mar 13 08:46:00 crc kubenswrapper[4876]: E0313 08:46:00.150563 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerName="extract-content" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.150652 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerName="extract-content" Mar 13 08:46:00 crc kubenswrapper[4876]: E0313 08:46:00.150732 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673c9f92-1c36-445c-8566-aba08dc8a304" containerName="container-00" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.150787 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="673c9f92-1c36-445c-8566-aba08dc8a304" containerName="container-00" Mar 13 08:46:00 crc kubenswrapper[4876]: E0313 08:46:00.150851 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerName="extract-utilities" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.150909 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerName="extract-utilities" Mar 13 08:46:00 crc kubenswrapper[4876]: E0313 08:46:00.150984 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerName="registry-server" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.151039 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerName="registry-server" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.151273 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="44011e32-c93d-44d2-93ae-5d8a3be8761e" containerName="registry-server" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.151353 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="673c9f92-1c36-445c-8566-aba08dc8a304" containerName="container-00" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.152065 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556526-cgtsw" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.154978 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.155753 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.155922 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.175853 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556526-cgtsw"] Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.207758 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8q6f\" (UniqueName: \"kubernetes.io/projected/b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c-kube-api-access-t8q6f\") pod \"auto-csr-approver-29556526-cgtsw\" (UID: \"b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c\") " pod="openshift-infra/auto-csr-approver-29556526-cgtsw" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.309728 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8q6f\" (UniqueName: \"kubernetes.io/projected/b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c-kube-api-access-t8q6f\") pod \"auto-csr-approver-29556526-cgtsw\" (UID: \"b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c\") " pod="openshift-infra/auto-csr-approver-29556526-cgtsw" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.339004 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8q6f\" (UniqueName: \"kubernetes.io/projected/b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c-kube-api-access-t8q6f\") pod \"auto-csr-approver-29556526-cgtsw\" (UID: \"b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c\") " pod="openshift-infra/auto-csr-approver-29556526-cgtsw" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.474104 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556526-cgtsw" Mar 13 08:46:00 crc kubenswrapper[4876]: I0313 08:46:00.914391 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556526-cgtsw"] Mar 13 08:46:01 crc kubenswrapper[4876]: I0313 08:46:01.783520 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556526-cgtsw" event={"ID":"b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c","Type":"ContainerStarted","Data":"edb28539b7d34b4b03765aeaa7db8ccee91eff57283f29dde216a87c30ce2d03"} Mar 13 08:46:02 crc kubenswrapper[4876]: I0313 08:46:02.794290 4876 generic.go:334] "Generic (PLEG): container finished" podID="b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c" containerID="4bb87886aeeb890284330a4a0a11684d110cd4d348a38ed6c25974f7a40e3f86" exitCode=0 Mar 13 08:46:02 crc kubenswrapper[4876]: I0313 08:46:02.794364 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556526-cgtsw" event={"ID":"b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c","Type":"ContainerDied","Data":"4bb87886aeeb890284330a4a0a11684d110cd4d348a38ed6c25974f7a40e3f86"} Mar 13 08:46:04 crc kubenswrapper[4876]: I0313 08:46:04.232037 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556526-cgtsw" Mar 13 08:46:04 crc kubenswrapper[4876]: I0313 08:46:04.319257 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8q6f\" (UniqueName: \"kubernetes.io/projected/b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c-kube-api-access-t8q6f\") pod \"b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c\" (UID: \"b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c\") " Mar 13 08:46:04 crc kubenswrapper[4876]: I0313 08:46:04.327732 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c-kube-api-access-t8q6f" (OuterVolumeSpecName: "kube-api-access-t8q6f") pod "b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c" (UID: "b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c"). InnerVolumeSpecName "kube-api-access-t8q6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:46:04 crc kubenswrapper[4876]: I0313 08:46:04.421161 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8q6f\" (UniqueName: \"kubernetes.io/projected/b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c-kube-api-access-t8q6f\") on node \"crc\" DevicePath \"\"" Mar 13 08:46:04 crc kubenswrapper[4876]: I0313 08:46:04.810947 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556526-cgtsw" event={"ID":"b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c","Type":"ContainerDied","Data":"edb28539b7d34b4b03765aeaa7db8ccee91eff57283f29dde216a87c30ce2d03"} Mar 13 08:46:04 crc kubenswrapper[4876]: I0313 08:46:04.810990 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edb28539b7d34b4b03765aeaa7db8ccee91eff57283f29dde216a87c30ce2d03" Mar 13 08:46:04 crc kubenswrapper[4876]: I0313 08:46:04.810995 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556526-cgtsw" Mar 13 08:46:05 crc kubenswrapper[4876]: I0313 08:46:05.312069 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556520-nm5cz"] Mar 13 08:46:05 crc kubenswrapper[4876]: I0313 08:46:05.331419 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556520-nm5cz"] Mar 13 08:46:06 crc kubenswrapper[4876]: I0313 08:46:06.365183 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8b86bd47d-2hv8s_65a5d87e-b9ef-4be3-ade6-46fa03282d7d/barbican-api/0.log" Mar 13 08:46:06 crc kubenswrapper[4876]: I0313 08:46:06.559951 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8b86bd47d-2hv8s_65a5d87e-b9ef-4be3-ade6-46fa03282d7d/barbican-api-log/0.log" Mar 13 08:46:06 crc kubenswrapper[4876]: I0313 08:46:06.633629 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5864c7c4b6-qqb8c_79f28ab7-204b-4e1c-be09-9e48b9ed5a75/barbican-keystone-listener/0.log" Mar 13 08:46:06 crc kubenswrapper[4876]: I0313 08:46:06.673901 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5864c7c4b6-qqb8c_79f28ab7-204b-4e1c-be09-9e48b9ed5a75/barbican-keystone-listener-log/0.log" Mar 13 08:46:06 crc kubenswrapper[4876]: I0313 08:46:06.850908 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5467d6f7c5-djg95_94630a52-4ec8-4ba9-966b-334ec46e9d25/barbican-worker/0.log" Mar 13 08:46:06 crc kubenswrapper[4876]: I0313 08:46:06.862158 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5467d6f7c5-djg95_94630a52-4ec8-4ba9-966b-334ec46e9d25/barbican-worker-log/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.046972 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7631eb4-cbcc-4df8-a4a8-46f00224c41e" path="/var/lib/kubelet/pods/e7631eb4-cbcc-4df8-a4a8-46f00224c41e/volumes" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.208689 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gn94z_0808c5a4-53f5-43e5-af39-bf78530e1d71/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.317903 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_27e28286-b51c-43cf-acbb-d612cae68530/ceilometer-central-agent/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.352923 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_27e28286-b51c-43cf-acbb-d612cae68530/proxy-httpd/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.399934 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_27e28286-b51c-43cf-acbb-d612cae68530/ceilometer-notification-agent/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.413805 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_27e28286-b51c-43cf-acbb-d612cae68530/sg-core/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.548446 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_19442bc3-f0a0-4077-8b11-823aeb39dc3f/cinder-api-log/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.649873 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_19442bc3-f0a0-4077-8b11-823aeb39dc3f/cinder-api/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.786794 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c3cb708-df89-4a10-ba84-b0c55efd2b61/probe/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.843548 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4c3cb708-df89-4a10-ba84-b0c55efd2b61/cinder-scheduler/0.log" Mar 13 08:46:07 crc kubenswrapper[4876]: I0313 08:46:07.960051 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-jkvx5_5bd6dd59-7427-4e76-8570-0a334476406c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.045739 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-pvxn4_5268b132-b205-4741-8680-781f73f3209a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.176418 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-vvnsb_6de4174c-e8e2-4058-a35b-c277eced33f9/init/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.387127 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-fwnht_0de3f5bc-8761-491e-9a81-9a03e796ca1b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.402614 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-vvnsb_6de4174c-e8e2-4058-a35b-c277eced33f9/init/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.458920 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-vvnsb_6de4174c-e8e2-4058-a35b-c277eced33f9/dnsmasq-dns/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.642003 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e4def3c5-fa14-4aed-94c0-3374552b56cf/glance-httpd/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.666520 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e4def3c5-fa14-4aed-94c0-3374552b56cf/glance-log/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.804801 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_dd9df755-7613-4abe-a9fc-661e6667eb63/glance-httpd/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.829173 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_dd9df755-7613-4abe-a9fc-661e6667eb63/glance-log/0.log" Mar 13 08:46:08 crc kubenswrapper[4876]: I0313 08:46:08.985063 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-764bff8b56-4l8lg_70b62196-7476-4c85-a2d7-2299da069989/horizon/0.log" Mar 13 08:46:09 crc kubenswrapper[4876]: I0313 08:46:09.277404 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xdclw_92d53dff-55bb-4953-96a9-647552c446e0/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:09 crc kubenswrapper[4876]: I0313 08:46:09.336832 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-764bff8b56-4l8lg_70b62196-7476-4c85-a2d7-2299da069989/horizon-log/0.log" Mar 13 08:46:09 crc kubenswrapper[4876]: I0313 08:46:09.445724 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4m5cw_fb7b397a-4ce3-4a83-b236-5baac12e2a15/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:09 crc kubenswrapper[4876]: I0313 08:46:09.629833 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-556bf8599b-szn97_b9195731-b1ca-4c6c-90f3-54549fa3a813/keystone-api/0.log" Mar 13 08:46:09 crc kubenswrapper[4876]: I0313 08:46:09.654087 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29556481-xrkdk_a91beee0-58e6-4d98-8000-e4e6bbee83a8/keystone-cron/0.log" Mar 13 08:46:09 crc kubenswrapper[4876]: I0313 08:46:09.856564 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_ff27b563-91dc-4be1-b0b8-98955850d2dd/kube-state-metrics/0.log" Mar 13 08:46:09 crc kubenswrapper[4876]: I0313 08:46:09.951076 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-v9b5k_240f0880-695b-44b4-ae52-2beacd34f2d0/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:10 crc kubenswrapper[4876]: I0313 08:46:10.280075 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5884ccb8d9-4wjtm_cfccefa5-a3f2-4513-a79b-372dd903ac2d/neutron-api/0.log" Mar 13 08:46:10 crc kubenswrapper[4876]: I0313 08:46:10.313479 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5884ccb8d9-4wjtm_cfccefa5-a3f2-4513-a79b-372dd903ac2d/neutron-httpd/0.log" Mar 13 08:46:10 crc kubenswrapper[4876]: I0313 08:46:10.636164 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jcshg_97951031-bd23-462e-83a8-be7068e1dfc8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:11 crc kubenswrapper[4876]: I0313 08:46:11.206147 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a77871ec-0f52-4d09-a3e2-664d14c9adbf/nova-cell0-conductor-conductor/0.log" Mar 13 08:46:11 crc kubenswrapper[4876]: I0313 08:46:11.252477 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cfd0230f-78c9-4fa3-a450-889afd1fb40a/nova-api-log/0.log" Mar 13 08:46:11 crc kubenswrapper[4876]: I0313 08:46:11.554889 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1f7187f5-a8b9-47ba-99bd-90a6a1cf5d83/nova-cell1-conductor-conductor/0.log" Mar 13 08:46:11 crc kubenswrapper[4876]: I0313 08:46:11.685996 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9b56d8da-0553-46dd-b2ab-26df263e1885/nova-cell1-novncproxy-novncproxy/0.log" Mar 13 08:46:11 crc kubenswrapper[4876]: I0313 08:46:11.690450 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cfd0230f-78c9-4fa3-a450-889afd1fb40a/nova-api-api/0.log" Mar 13 08:46:11 crc kubenswrapper[4876]: I0313 08:46:11.802416 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-2xhz4_7b648455-e602-439e-b806-743233acb285/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:12 crc kubenswrapper[4876]: I0313 08:46:12.026521 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_33193662-0ae1-4e6d-bcb8-d093ee42b647/nova-metadata-log/0.log" Mar 13 08:46:12 crc kubenswrapper[4876]: I0313 08:46:12.358052 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e8c36ecf-187d-4388-8d6f-626fe6470eb0/mysql-bootstrap/0.log" Mar 13 08:46:12 crc kubenswrapper[4876]: I0313 08:46:12.485688 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d51bcb66-a417-4369-94df-a59aedd82d1c/nova-scheduler-scheduler/0.log" Mar 13 08:46:12 crc kubenswrapper[4876]: I0313 08:46:12.593611 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e8c36ecf-187d-4388-8d6f-626fe6470eb0/mysql-bootstrap/0.log" Mar 13 08:46:12 crc kubenswrapper[4876]: I0313 08:46:12.663484 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e8c36ecf-187d-4388-8d6f-626fe6470eb0/galera/0.log" Mar 13 08:46:12 crc kubenswrapper[4876]: I0313 08:46:12.805214 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7e32caff-0034-47f2-aa5b-917607dd44b6/mysql-bootstrap/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.062418 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7e32caff-0034-47f2-aa5b-917607dd44b6/mysql-bootstrap/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.065541 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7e32caff-0034-47f2-aa5b-917607dd44b6/galera/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.257487 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_02e40021-e864-4c53-9522-40fe5719abde/openstackclient/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.341651 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rgh2p_d0dbe7f9-9cce-4b19-86d6-1b445c996d95/openstack-network-exporter/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.455572 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_33193662-0ae1-4e6d-bcb8-d093ee42b647/nova-metadata-metadata/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.546527 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tb4jh_23f73ce7-85a0-410c-a3f3-f1df187b509e/ovsdb-server-init/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.726134 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tb4jh_23f73ce7-85a0-410c-a3f3-f1df187b509e/ovsdb-server/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.737700 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tb4jh_23f73ce7-85a0-410c-a3f3-f1df187b509e/ovs-vswitchd/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.753522 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tb4jh_23f73ce7-85a0-410c-a3f3-f1df187b509e/ovsdb-server-init/0.log" Mar 13 08:46:13 crc kubenswrapper[4876]: I0313 08:46:13.987657 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xrn6k_3b55481b-c5a5-4602-8f66-c3b002134edb/ovn-controller/0.log" Mar 13 08:46:14 crc kubenswrapper[4876]: I0313 08:46:14.048134 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hrnbj_848b0256-b349-49c3-a007-20c0da303ad8/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:14 crc kubenswrapper[4876]: I0313 08:46:14.351040 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_76257d47-12f0-4e22-a3d7-856367d8a08b/openstack-network-exporter/0.log" Mar 13 08:46:14 crc kubenswrapper[4876]: I0313 08:46:14.430727 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_76257d47-12f0-4e22-a3d7-856367d8a08b/ovn-northd/0.log" Mar 13 08:46:14 crc kubenswrapper[4876]: I0313 08:46:14.600888 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_88e4b876-8c78-4d88-945b-bd10555503d2/ovsdbserver-nb/0.log" Mar 13 08:46:14 crc kubenswrapper[4876]: I0313 08:46:14.625898 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_88e4b876-8c78-4d88-945b-bd10555503d2/openstack-network-exporter/0.log" Mar 13 08:46:14 crc kubenswrapper[4876]: I0313 08:46:14.788351 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37b39bdf-9f1c-4f03-88e5-6abd3949fc7c/openstack-network-exporter/0.log" Mar 13 08:46:14 crc kubenswrapper[4876]: I0313 08:46:14.866873 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37b39bdf-9f1c-4f03-88e5-6abd3949fc7c/ovsdbserver-sb/0.log" Mar 13 08:46:14 crc kubenswrapper[4876]: I0313 08:46:14.981411 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6dbd6d4d74-7zn9n_9ab4e2aa-c297-4979-9322-78f14c7b36f8/placement-api/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.086568 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147c88b9-db25-435e-9575-edb4efa342e7/setup-container/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.127638 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6dbd6d4d74-7zn9n_9ab4e2aa-c297-4979-9322-78f14c7b36f8/placement-log/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.348609 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_79aa8084-23ef-4730-948d-1b0d6ca25251/setup-container/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.358614 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147c88b9-db25-435e-9575-edb4efa342e7/rabbitmq/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.379120 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147c88b9-db25-435e-9575-edb4efa342e7/setup-container/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.643370 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_79aa8084-23ef-4730-948d-1b0d6ca25251/setup-container/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.646943 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_79aa8084-23ef-4730-948d-1b0d6ca25251/rabbitmq/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.650265 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4nnsf_8caf73f8-76f3-4fa6-bfbc-9977576c8d18/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:15 crc kubenswrapper[4876]: I0313 08:46:15.861488 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-cm4qh_1f20780c-14f1-4165-95a5-15b604018e4b/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.083910 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-n7thg_61be635e-fed6-4b19-9095-cbaa3720b693/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.251440 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-hj2tl_f87ce9a3-0a04-4b72-bd4f-88a5538c99c1/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.353676 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-6q2l4_ca0807c6-78af-45cf-994d-85563d8089ac/ssh-known-hosts-edpm-deployment/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.528832 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b9c5d6d4c-7lxwx_62fd9eb3-6af0-4bc9-bc57-65c8b03eb467/proxy-server/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.593612 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-pbnh2_e5294658-a1f0-4340-9d61-c8008a37b5e9/swift-ring-rebalance/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.660047 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b9c5d6d4c-7lxwx_62fd9eb3-6af0-4bc9-bc57-65c8b03eb467/proxy-httpd/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.758609 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/account-auditor/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.892760 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/account-reaper/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.900086 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/account-replicator/0.log" Mar 13 08:46:16 crc kubenswrapper[4876]: I0313 08:46:16.984184 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/account-server/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.001730 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/container-auditor/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.105792 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/container-server/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.176181 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/container-replicator/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.237985 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/container-updater/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.250756 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-auditor/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.332774 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-expirer/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.431395 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-replicator/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.432614 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-server/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.466965 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/object-updater/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.551989 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/rsync/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.830342 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77efd2bc-d283-4ab6-bd5f-b346b431ea09/swift-recon-cron/0.log" Mar 13 08:46:17 crc kubenswrapper[4876]: I0313 08:46:17.995132 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-rd6cx_18be12a3-56e2-4b39-8c8f-784a7cafed2b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:18 crc kubenswrapper[4876]: I0313 08:46:18.187457 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_009d8699-15dd-4c89-bd6c-5e7a936e1b59/tempest-tests-tempest-tests-runner/0.log" Mar 13 08:46:18 crc kubenswrapper[4876]: I0313 08:46:18.260739 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_018a8dfd-03af-4cc5-9689-78f7f3472d0d/test-operator-logs-container/0.log" Mar 13 08:46:18 crc kubenswrapper[4876]: I0313 08:46:18.366155 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qljdp_5b3fa1e6-19bc-4852-bf70-69051aa084d6/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 08:46:26 crc kubenswrapper[4876]: I0313 08:46:26.806418 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e221cd11-36ee-4c17-a26d-e95b75ed5682/memcached/0.log" Mar 13 08:46:29 crc kubenswrapper[4876]: I0313 08:46:29.942557 4876 scope.go:117] "RemoveContainer" containerID="2cf7f22e2f77c14f865113d9307e85633f40a1b93af8c479e357b60cf251b911" Mar 13 08:46:44 crc kubenswrapper[4876]: I0313 08:46:44.334453 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/util/0.log" Mar 13 08:46:44 crc kubenswrapper[4876]: I0313 08:46:44.607973 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/util/0.log" Mar 13 08:46:44 crc kubenswrapper[4876]: I0313 08:46:44.627103 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/pull/0.log" Mar 13 08:46:44 crc kubenswrapper[4876]: I0313 08:46:44.630633 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/pull/0.log" Mar 13 08:46:45 crc kubenswrapper[4876]: I0313 08:46:45.204039 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/util/0.log" Mar 13 08:46:45 crc kubenswrapper[4876]: I0313 08:46:45.205720 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/extract/0.log" Mar 13 08:46:45 crc kubenswrapper[4876]: I0313 08:46:45.236463 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c53a0f00d944ccae55960377341d5d563a14e47622a37a940c96d3cce86tj2x_28ee98e2-f56e-4050-ae8a-6c4fe84eda60/pull/0.log" Mar 13 08:46:45 crc kubenswrapper[4876]: I0313 08:46:45.614877 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-897gd_68223df4-94ca-47fe-afbc-93c77725ba2a/manager/0.log" Mar 13 08:46:45 crc kubenswrapper[4876]: I0313 08:46:45.969108 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-zlt8c_151ed757-e99e-4e7e-bfaa-7dec0ef8b038/manager/0.log" Mar 13 08:46:46 crc kubenswrapper[4876]: I0313 08:46:46.161910 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-jcjfd_8c04b15f-3ca3-4875-9ea4-45244808af5f/manager/0.log" Mar 13 08:46:46 crc kubenswrapper[4876]: I0313 08:46:46.331899 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-c6bfj_ff48d75e-19f2-4a48-bcd6-3855b494fdf1/manager/0.log" Mar 13 08:46:47 crc kubenswrapper[4876]: I0313 08:46:47.182022 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6bbb499bbc-n5stb_3ff4d1b1-aae8-4900-a317-164ea72a78b2/manager/0.log" Mar 13 08:46:47 crc kubenswrapper[4876]: I0313 08:46:47.342814 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5995f4446f-lcwhv_722c5d70-ed42-4b75-8993-3b3643d44d1a/manager/0.log" Mar 13 08:46:47 crc kubenswrapper[4876]: I0313 08:46:47.606698 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-w45l7_8d21a99d-11a5-4b5f-a44d-3f9cce84b82a/manager/0.log" Mar 13 08:46:47 crc kubenswrapper[4876]: I0313 08:46:47.614652 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-kmhm8_33e3356e-590a-4b31-990f-4ecda38efb0a/manager/0.log" Mar 13 08:46:47 crc kubenswrapper[4876]: I0313 08:46:47.813911 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-68f45f9d9f-bp9bt_98ac6099-e731-43c8-8537-8095f1bfce09/manager/0.log" Mar 13 08:46:47 crc kubenswrapper[4876]: I0313 08:46:47.910741 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-658d4cdd5-v7kpw_6aeda87b-2ede-4dcb-9234-bfc8837f4ea8/manager/0.log" Mar 13 08:46:48 crc kubenswrapper[4876]: I0313 08:46:48.167371 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-lnb48_4284fe2d-d1bb-49f6-b555-0f46419ad11b/manager/0.log" Mar 13 08:46:48 crc kubenswrapper[4876]: I0313 08:46:48.402907 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-m2xqm_3de3f00b-74de-4328-86e9-9f86c7277deb/manager/0.log" Mar 13 08:46:48 crc kubenswrapper[4876]: I0313 08:46:48.426469 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-569cc54c5-pfswh_ef934388-2da5-47fe-8679-ec520946de1b/manager/0.log" Mar 13 08:46:48 crc kubenswrapper[4876]: I0313 08:46:48.546127 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7m6k9q_256daf30-6602-4a5d-8a4b-0f96fb5450b2/manager/0.log" Mar 13 08:46:48 crc kubenswrapper[4876]: I0313 08:46:48.728559 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-578cd8b599-6vlp7_995e15f2-9906-40df-a498-832366249bf9/operator/0.log" Mar 13 08:46:48 crc kubenswrapper[4876]: I0313 08:46:48.938328 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tfbgk_ec8ee7ae-49f7-473f-b432-9c2395e93115/registry-server/0.log" Mar 13 08:46:49 crc kubenswrapper[4876]: I0313 08:46:49.145682 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-d5d5r_dacb8b62-d4f3-43af-88b2-26fbc725f50f/manager/0.log" Mar 13 08:46:49 crc kubenswrapper[4876]: I0313 08:46:49.252842 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-m7zdz_a9a4fe04-37f1-47ad-ab32-e983f13df515/manager/0.log" Mar 13 08:46:49 crc kubenswrapper[4876]: I0313 08:46:49.358986 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-tzshq_8a6e1029-8d6f-4698-b92b-d0cd17bda62d/operator/0.log" Mar 13 08:46:49 crc kubenswrapper[4876]: I0313 08:46:49.559551 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-677c674df7-hgmsc_e7f3a695-6649-482f-94d9-238966224ba8/manager/0.log" Mar 13 08:46:49 crc kubenswrapper[4876]: I0313 08:46:49.793875 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cd66dbd4b-jt5rp_a314a819-a02b-4e9f-a996-58bed1d2e6be/manager/0.log" Mar 13 08:46:49 crc kubenswrapper[4876]: I0313 08:46:49.871489 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-jztw5_78f4bc78-4ca4-4bb1-a7d5-99627fd5e02e/manager/0.log" Mar 13 08:46:50 crc kubenswrapper[4876]: I0313 08:46:50.005766 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6dd88c6f67-cm65k_3aefe2ee-3d59-46e9-9a6e-4f576566a49c/manager/0.log" Mar 13 08:46:50 crc kubenswrapper[4876]: I0313 08:46:50.455914 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74f6d57fdc-mfkh6_bad784ad-2cd8-4f50-832f-0e57d06249d6/manager/0.log" Mar 13 08:46:53 crc kubenswrapper[4876]: I0313 08:46:53.993790 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-6slg2_4cfaf185-ed2d-4c36-9d38-fcf4e74905ec/manager/0.log" Mar 13 08:47:10 crc kubenswrapper[4876]: I0313 08:47:10.973928 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-hln48_8c6d92f4-c514-48e4-b1da-61546f07d171/control-plane-machine-set-operator/0.log" Mar 13 08:47:11 crc kubenswrapper[4876]: I0313 08:47:11.123076 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xvwj2_418bec94-3753-4d76-b2a8-24b04f39c820/kube-rbac-proxy/0.log" Mar 13 08:47:11 crc kubenswrapper[4876]: I0313 08:47:11.186958 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xvwj2_418bec94-3753-4d76-b2a8-24b04f39c820/machine-api-operator/0.log" Mar 13 08:47:24 crc kubenswrapper[4876]: I0313 08:47:24.102714 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-552r7_bbd6238b-a711-47f6-bf54-ae2b35e8c79d/cert-manager-controller/0.log" Mar 13 08:47:24 crc kubenswrapper[4876]: I0313 08:47:24.212461 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-2c8kw_037a0101-dff7-49a6-9974-a5b88c2cb7f1/cert-manager-cainjector/0.log" Mar 13 08:47:24 crc kubenswrapper[4876]: I0313 08:47:24.297279 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-jm7bb_d0aa3593-da37-4da6-ae1f-890cb901c029/cert-manager-webhook/0.log" Mar 13 08:47:38 crc kubenswrapper[4876]: I0313 08:47:38.492645 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-62blj_be89d0d3-8483-4439-b346-c50bd885b15c/nmstate-console-plugin/0.log" Mar 13 08:47:38 crc kubenswrapper[4876]: I0313 08:47:38.669528 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-v9cql_6aa5c930-3567-401a-830f-c30ca123b57a/nmstate-handler/0.log" Mar 13 08:47:38 crc kubenswrapper[4876]: I0313 08:47:38.744798 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-d2mlg_8f70b2a8-97e0-476b-91d2-e5c386d8a14f/kube-rbac-proxy/0.log" Mar 13 08:47:38 crc kubenswrapper[4876]: I0313 08:47:38.822084 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-d2mlg_8f70b2a8-97e0-476b-91d2-e5c386d8a14f/nmstate-metrics/0.log" Mar 13 08:47:38 crc kubenswrapper[4876]: I0313 08:47:38.939013 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-5gv4s_54e0cdfc-3234-4763-bc49-6a6e7a537c74/nmstate-operator/0.log" Mar 13 08:47:39 crc kubenswrapper[4876]: I0313 08:47:39.032164 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-74bqr_e506e0e1-25c2-4be3-ad95-5f9029a960e3/nmstate-webhook/0.log" Mar 13 08:47:54 crc kubenswrapper[4876]: I0313 08:47:54.608163 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:47:54 crc kubenswrapper[4876]: I0313 08:47:54.609174 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.175228 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556528-6rz8d"] Mar 13 08:48:00 crc kubenswrapper[4876]: E0313 08:48:00.179949 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c" containerName="oc" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.179970 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c" containerName="oc" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.180148 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c" containerName="oc" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.180878 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556528-6rz8d" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.183698 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.184126 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.184354 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.192114 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556528-6rz8d"] Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.231087 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5jrx\" (UniqueName: \"kubernetes.io/projected/1acea7c3-7bed-454e-bfc5-90ea76fd4a51-kube-api-access-z5jrx\") pod \"auto-csr-approver-29556528-6rz8d\" (UID: \"1acea7c3-7bed-454e-bfc5-90ea76fd4a51\") " pod="openshift-infra/auto-csr-approver-29556528-6rz8d" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.334019 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5jrx\" (UniqueName: \"kubernetes.io/projected/1acea7c3-7bed-454e-bfc5-90ea76fd4a51-kube-api-access-z5jrx\") pod \"auto-csr-approver-29556528-6rz8d\" (UID: \"1acea7c3-7bed-454e-bfc5-90ea76fd4a51\") " pod="openshift-infra/auto-csr-approver-29556528-6rz8d" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.366803 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5jrx\" (UniqueName: \"kubernetes.io/projected/1acea7c3-7bed-454e-bfc5-90ea76fd4a51-kube-api-access-z5jrx\") pod \"auto-csr-approver-29556528-6rz8d\" (UID: \"1acea7c3-7bed-454e-bfc5-90ea76fd4a51\") " pod="openshift-infra/auto-csr-approver-29556528-6rz8d" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.504042 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556528-6rz8d" Mar 13 08:48:00 crc kubenswrapper[4876]: I0313 08:48:00.972954 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556528-6rz8d"] Mar 13 08:48:01 crc kubenswrapper[4876]: I0313 08:48:01.937494 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556528-6rz8d" event={"ID":"1acea7c3-7bed-454e-bfc5-90ea76fd4a51","Type":"ContainerStarted","Data":"9fff350ca7f986ca9f648aba2cb1e5d9ec59c0d3ca4538e5453162db2c39e394"} Mar 13 08:48:02 crc kubenswrapper[4876]: I0313 08:48:02.949637 4876 generic.go:334] "Generic (PLEG): container finished" podID="1acea7c3-7bed-454e-bfc5-90ea76fd4a51" containerID="0af5c7cea6d8d1cdc7d3131732b0fc2ee1f360021c86c4cec9642b2f850e3034" exitCode=0 Mar 13 08:48:02 crc kubenswrapper[4876]: I0313 08:48:02.949693 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556528-6rz8d" event={"ID":"1acea7c3-7bed-454e-bfc5-90ea76fd4a51","Type":"ContainerDied","Data":"0af5c7cea6d8d1cdc7d3131732b0fc2ee1f360021c86c4cec9642b2f850e3034"} Mar 13 08:48:04 crc kubenswrapper[4876]: I0313 08:48:04.420345 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556528-6rz8d" Mar 13 08:48:04 crc kubenswrapper[4876]: I0313 08:48:04.520164 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5jrx\" (UniqueName: \"kubernetes.io/projected/1acea7c3-7bed-454e-bfc5-90ea76fd4a51-kube-api-access-z5jrx\") pod \"1acea7c3-7bed-454e-bfc5-90ea76fd4a51\" (UID: \"1acea7c3-7bed-454e-bfc5-90ea76fd4a51\") " Mar 13 08:48:04 crc kubenswrapper[4876]: I0313 08:48:04.526454 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1acea7c3-7bed-454e-bfc5-90ea76fd4a51-kube-api-access-z5jrx" (OuterVolumeSpecName: "kube-api-access-z5jrx") pod "1acea7c3-7bed-454e-bfc5-90ea76fd4a51" (UID: "1acea7c3-7bed-454e-bfc5-90ea76fd4a51"). InnerVolumeSpecName "kube-api-access-z5jrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:48:04 crc kubenswrapper[4876]: I0313 08:48:04.622082 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5jrx\" (UniqueName: \"kubernetes.io/projected/1acea7c3-7bed-454e-bfc5-90ea76fd4a51-kube-api-access-z5jrx\") on node \"crc\" DevicePath \"\"" Mar 13 08:48:04 crc kubenswrapper[4876]: I0313 08:48:04.974118 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556528-6rz8d" event={"ID":"1acea7c3-7bed-454e-bfc5-90ea76fd4a51","Type":"ContainerDied","Data":"9fff350ca7f986ca9f648aba2cb1e5d9ec59c0d3ca4538e5453162db2c39e394"} Mar 13 08:48:04 crc kubenswrapper[4876]: I0313 08:48:04.974200 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556528-6rz8d" Mar 13 08:48:04 crc kubenswrapper[4876]: I0313 08:48:04.974203 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fff350ca7f986ca9f648aba2cb1e5d9ec59c0d3ca4538e5453162db2c39e394" Mar 13 08:48:05 crc kubenswrapper[4876]: I0313 08:48:05.507833 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556522-whtwn"] Mar 13 08:48:05 crc kubenswrapper[4876]: I0313 08:48:05.523273 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556522-whtwn"] Mar 13 08:48:07 crc kubenswrapper[4876]: I0313 08:48:07.053636 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb9e0427-5268-4b70-bec0-c1eed76b6554" path="/var/lib/kubelet/pods/bb9e0427-5268-4b70-bec0-c1eed76b6554/volumes" Mar 13 08:48:09 crc kubenswrapper[4876]: I0313 08:48:09.884037 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-bmq26_97a22a43-697e-4c6a-b800-489008955cbe/kube-rbac-proxy/0.log" Mar 13 08:48:09 crc kubenswrapper[4876]: I0313 08:48:09.986476 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-bmq26_97a22a43-697e-4c6a-b800-489008955cbe/controller/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.092664 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-frr-files/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.317729 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-frr-files/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.331200 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-reloader/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.348368 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-reloader/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.372075 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-metrics/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.538438 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-frr-files/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.550094 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-metrics/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.563143 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-reloader/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.620534 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-metrics/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.748015 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-reloader/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.749857 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-frr-files/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.781662 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/cp-metrics/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.840639 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/controller/0.log" Mar 13 08:48:10 crc kubenswrapper[4876]: I0313 08:48:10.997435 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/frr-metrics/0.log" Mar 13 08:48:11 crc kubenswrapper[4876]: I0313 08:48:11.047475 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/kube-rbac-proxy/0.log" Mar 13 08:48:11 crc kubenswrapper[4876]: I0313 08:48:11.086303 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/kube-rbac-proxy-frr/0.log" Mar 13 08:48:11 crc kubenswrapper[4876]: I0313 08:48:11.206759 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/reloader/0.log" Mar 13 08:48:11 crc kubenswrapper[4876]: I0313 08:48:11.335702 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-6vvzq_91b6f5ca-3921-446a-a57a-de264ca63118/frr-k8s-webhook-server/0.log" Mar 13 08:48:11 crc kubenswrapper[4876]: I0313 08:48:11.512156 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5f79f8644-vxb58_122ad6c8-564c-4aff-8dc5-6515b0e9b5cb/manager/0.log" Mar 13 08:48:11 crc kubenswrapper[4876]: I0313 08:48:11.659999 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-679dfbf9b7-f5wpn_e918dae3-24c9-4e8f-8f89-f4070580eeff/webhook-server/0.log" Mar 13 08:48:11 crc kubenswrapper[4876]: I0313 08:48:11.870190 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nb74f_65146c95-889a-4030-b7d3-3a1810b158de/kube-rbac-proxy/0.log" Mar 13 08:48:12 crc kubenswrapper[4876]: I0313 08:48:12.429834 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nb74f_65146c95-889a-4030-b7d3-3a1810b158de/speaker/0.log" Mar 13 08:48:12 crc kubenswrapper[4876]: I0313 08:48:12.825571 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hmszn_0979a634-546c-441c-8438-51ab6723d891/frr/0.log" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.482872 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fx8kd"] Mar 13 08:48:23 crc kubenswrapper[4876]: E0313 08:48:23.484374 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1acea7c3-7bed-454e-bfc5-90ea76fd4a51" containerName="oc" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.484395 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1acea7c3-7bed-454e-bfc5-90ea76fd4a51" containerName="oc" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.484670 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1acea7c3-7bed-454e-bfc5-90ea76fd4a51" containerName="oc" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.486514 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.506926 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx8kd"] Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.608512 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28t8w\" (UniqueName: \"kubernetes.io/projected/7378eb49-c592-449a-82d3-07f9e7469b78-kube-api-access-28t8w\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.608613 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-catalog-content\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.608984 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-utilities\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.711066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28t8w\" (UniqueName: \"kubernetes.io/projected/7378eb49-c592-449a-82d3-07f9e7469b78-kube-api-access-28t8w\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.711215 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-catalog-content\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.711368 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-utilities\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.711887 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-catalog-content\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.712063 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-utilities\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.738377 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28t8w\" (UniqueName: \"kubernetes.io/projected/7378eb49-c592-449a-82d3-07f9e7469b78-kube-api-access-28t8w\") pod \"redhat-marketplace-fx8kd\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:23 crc kubenswrapper[4876]: I0313 08:48:23.812831 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:24 crc kubenswrapper[4876]: I0313 08:48:24.314102 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx8kd"] Mar 13 08:48:24 crc kubenswrapper[4876]: W0313 08:48:24.320470 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7378eb49_c592_449a_82d3_07f9e7469b78.slice/crio-36a472438888a72a74e26fd856b752adf4447da3b5aeb52b279764e1ed3b70be WatchSource:0}: Error finding container 36a472438888a72a74e26fd856b752adf4447da3b5aeb52b279764e1ed3b70be: Status 404 returned error can't find the container with id 36a472438888a72a74e26fd856b752adf4447da3b5aeb52b279764e1ed3b70be Mar 13 08:48:24 crc kubenswrapper[4876]: I0313 08:48:24.608085 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:48:24 crc kubenswrapper[4876]: I0313 08:48:24.608464 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:48:25 crc kubenswrapper[4876]: I0313 08:48:25.197378 4876 generic.go:334] "Generic (PLEG): container finished" podID="7378eb49-c592-449a-82d3-07f9e7469b78" containerID="8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f" exitCode=0 Mar 13 08:48:25 crc kubenswrapper[4876]: I0313 08:48:25.197462 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx8kd" event={"ID":"7378eb49-c592-449a-82d3-07f9e7469b78","Type":"ContainerDied","Data":"8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f"} Mar 13 08:48:25 crc kubenswrapper[4876]: I0313 08:48:25.197536 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx8kd" event={"ID":"7378eb49-c592-449a-82d3-07f9e7469b78","Type":"ContainerStarted","Data":"36a472438888a72a74e26fd856b752adf4447da3b5aeb52b279764e1ed3b70be"} Mar 13 08:48:26 crc kubenswrapper[4876]: I0313 08:48:26.209879 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx8kd" event={"ID":"7378eb49-c592-449a-82d3-07f9e7469b78","Type":"ContainerStarted","Data":"60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161"} Mar 13 08:48:27 crc kubenswrapper[4876]: I0313 08:48:27.220116 4876 generic.go:334] "Generic (PLEG): container finished" podID="7378eb49-c592-449a-82d3-07f9e7469b78" containerID="60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161" exitCode=0 Mar 13 08:48:27 crc kubenswrapper[4876]: I0313 08:48:27.220156 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx8kd" event={"ID":"7378eb49-c592-449a-82d3-07f9e7469b78","Type":"ContainerDied","Data":"60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161"} Mar 13 08:48:27 crc kubenswrapper[4876]: I0313 08:48:27.496628 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/util/0.log" Mar 13 08:48:27 crc kubenswrapper[4876]: I0313 08:48:27.712671 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/pull/0.log" Mar 13 08:48:27 crc kubenswrapper[4876]: I0313 08:48:27.715442 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/util/0.log" Mar 13 08:48:27 crc kubenswrapper[4876]: I0313 08:48:27.771291 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/pull/0.log" Mar 13 08:48:27 crc kubenswrapper[4876]: I0313 08:48:27.992196 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/extract/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.051771 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/pull/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.103116 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874tt4mt_ea82483b-90c0-4edd-b51c-38d2ae56de9e/util/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.186014 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/util/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.230001 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx8kd" event={"ID":"7378eb49-c592-449a-82d3-07f9e7469b78","Type":"ContainerStarted","Data":"a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba"} Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.252362 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fx8kd" podStartSLOduration=2.778123128 podStartE2EDuration="5.252342022s" podCreationTimestamp="2026-03-13 08:48:23 +0000 UTC" firstStartedPulling="2026-03-13 08:48:25.199467702 +0000 UTC m=+4164.870246684" lastFinishedPulling="2026-03-13 08:48:27.673686606 +0000 UTC m=+4167.344465578" observedRunningTime="2026-03-13 08:48:28.246955374 +0000 UTC m=+4167.917734366" watchObservedRunningTime="2026-03-13 08:48:28.252342022 +0000 UTC m=+4167.923121004" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.481705 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/util/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.502427 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/pull/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.513659 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/pull/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.667946 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/util/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.702796 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/extract/0.log" Mar 13 08:48:28 crc kubenswrapper[4876]: I0313 08:48:28.738983 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1p5f7c_489dadce-846b-4338-b396-1c80e256b2f8/pull/0.log" Mar 13 08:48:29 crc kubenswrapper[4876]: I0313 08:48:29.041770 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-utilities/0.log" Mar 13 08:48:29 crc kubenswrapper[4876]: I0313 08:48:29.183670 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-content/0.log" Mar 13 08:48:29 crc kubenswrapper[4876]: I0313 08:48:29.192918 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-content/0.log" Mar 13 08:48:29 crc kubenswrapper[4876]: I0313 08:48:29.232855 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-utilities/0.log" Mar 13 08:48:29 crc kubenswrapper[4876]: I0313 08:48:29.974700 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-content/0.log" Mar 13 08:48:29 crc kubenswrapper[4876]: I0313 08:48:29.978761 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/extract-utilities/0.log" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.097197 4876 scope.go:117] "RemoveContainer" containerID="653e0ba7e8bd41694fd2d300705bc2f149c649dd89711b73d929467fd6d025e6" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.215014 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-utilities/0.log" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.506916 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-content/0.log" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.522522 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-utilities/0.log" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.618194 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8rcs9_0e320bf9-7cd4-423f-af2a-9e00e602bcff/registry-server/0.log" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.622985 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-content/0.log" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.723848 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-content/0.log" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.757222 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/extract-utilities/0.log" Mar 13 08:48:30 crc kubenswrapper[4876]: I0313 08:48:30.969413 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gzmpk_a0adab6c-7942-4756-829f-3b7a6569f8cb/marketplace-operator/0.log" Mar 13 08:48:31 crc kubenswrapper[4876]: I0313 08:48:31.139855 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fx8kd_7378eb49-c592-449a-82d3-07f9e7469b78/extract-utilities/0.log" Mar 13 08:48:31 crc kubenswrapper[4876]: I0313 08:48:31.349026 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pxfhc_add82187-d01b-4902-8f12-ced45327d1f9/registry-server/0.log" Mar 13 08:48:31 crc kubenswrapper[4876]: I0313 08:48:31.355035 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fx8kd_7378eb49-c592-449a-82d3-07f9e7469b78/extract-utilities/0.log" Mar 13 08:48:31 crc kubenswrapper[4876]: I0313 08:48:31.360929 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fx8kd_7378eb49-c592-449a-82d3-07f9e7469b78/extract-content/0.log" Mar 13 08:48:31 crc kubenswrapper[4876]: I0313 08:48:31.389845 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fx8kd_7378eb49-c592-449a-82d3-07f9e7469b78/extract-content/0.log" Mar 13 08:48:32 crc kubenswrapper[4876]: I0313 08:48:32.413039 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fx8kd_7378eb49-c592-449a-82d3-07f9e7469b78/extract-utilities/0.log" Mar 13 08:48:32 crc kubenswrapper[4876]: I0313 08:48:32.432797 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fx8kd_7378eb49-c592-449a-82d3-07f9e7469b78/extract-content/0.log" Mar 13 08:48:32 crc kubenswrapper[4876]: I0313 08:48:32.442428 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-utilities/0.log" Mar 13 08:48:32 crc kubenswrapper[4876]: I0313 08:48:32.451852 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fx8kd_7378eb49-c592-449a-82d3-07f9e7469b78/registry-server/0.log" Mar 13 08:48:32 crc kubenswrapper[4876]: I0313 08:48:32.657224 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-content/0.log" Mar 13 08:48:32 crc kubenswrapper[4876]: I0313 08:48:32.695024 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-utilities/0.log" Mar 13 08:48:32 crc kubenswrapper[4876]: I0313 08:48:32.753025 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-content/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.136111 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-utilities/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.156780 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/extract-content/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.227556 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-utilities/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.308538 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gkkq5_cccd9f80-f78e-4e2b-a101-d85fa5b8a942/registry-server/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.410260 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-content/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.445652 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-content/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.448267 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-utilities/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.633226 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-utilities/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.640410 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/extract-content/0.log" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.813629 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.814049 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:33 crc kubenswrapper[4876]: I0313 08:48:33.880773 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:34 crc kubenswrapper[4876]: I0313 08:48:34.158694 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zvfb8_729a14e7-a89b-4e7c-a8b5-a4002519d6a3/registry-server/0.log" Mar 13 08:48:34 crc kubenswrapper[4876]: I0313 08:48:34.343721 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:34 crc kubenswrapper[4876]: I0313 08:48:34.398492 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx8kd"] Mar 13 08:48:36 crc kubenswrapper[4876]: I0313 08:48:36.320933 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fx8kd" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" containerName="registry-server" containerID="cri-o://a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba" gracePeriod=2 Mar 13 08:48:36 crc kubenswrapper[4876]: I0313 08:48:36.829775 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:36 crc kubenswrapper[4876]: I0313 08:48:36.915901 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-utilities\") pod \"7378eb49-c592-449a-82d3-07f9e7469b78\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " Mar 13 08:48:36 crc kubenswrapper[4876]: I0313 08:48:36.916209 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-catalog-content\") pod \"7378eb49-c592-449a-82d3-07f9e7469b78\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " Mar 13 08:48:36 crc kubenswrapper[4876]: I0313 08:48:36.916321 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28t8w\" (UniqueName: \"kubernetes.io/projected/7378eb49-c592-449a-82d3-07f9e7469b78-kube-api-access-28t8w\") pod \"7378eb49-c592-449a-82d3-07f9e7469b78\" (UID: \"7378eb49-c592-449a-82d3-07f9e7469b78\") " Mar 13 08:48:36 crc kubenswrapper[4876]: I0313 08:48:36.917792 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-utilities" (OuterVolumeSpecName: "utilities") pod "7378eb49-c592-449a-82d3-07f9e7469b78" (UID: "7378eb49-c592-449a-82d3-07f9e7469b78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:48:36 crc kubenswrapper[4876]: I0313 08:48:36.924403 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7378eb49-c592-449a-82d3-07f9e7469b78-kube-api-access-28t8w" (OuterVolumeSpecName: "kube-api-access-28t8w") pod "7378eb49-c592-449a-82d3-07f9e7469b78" (UID: "7378eb49-c592-449a-82d3-07f9e7469b78"). InnerVolumeSpecName "kube-api-access-28t8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:48:36 crc kubenswrapper[4876]: I0313 08:48:36.974298 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7378eb49-c592-449a-82d3-07f9e7469b78" (UID: "7378eb49-c592-449a-82d3-07f9e7469b78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.018895 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28t8w\" (UniqueName: \"kubernetes.io/projected/7378eb49-c592-449a-82d3-07f9e7469b78-kube-api-access-28t8w\") on node \"crc\" DevicePath \"\"" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.018932 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.018943 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7378eb49-c592-449a-82d3-07f9e7469b78-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.331864 4876 generic.go:334] "Generic (PLEG): container finished" podID="7378eb49-c592-449a-82d3-07f9e7469b78" containerID="a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba" exitCode=0 Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.331916 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fx8kd" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.331937 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx8kd" event={"ID":"7378eb49-c592-449a-82d3-07f9e7469b78","Type":"ContainerDied","Data":"a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba"} Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.332283 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx8kd" event={"ID":"7378eb49-c592-449a-82d3-07f9e7469b78","Type":"ContainerDied","Data":"36a472438888a72a74e26fd856b752adf4447da3b5aeb52b279764e1ed3b70be"} Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.332307 4876 scope.go:117] "RemoveContainer" containerID="a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.350113 4876 scope.go:117] "RemoveContainer" containerID="60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.360399 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx8kd"] Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.371983 4876 scope.go:117] "RemoveContainer" containerID="8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.376497 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx8kd"] Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.407371 4876 scope.go:117] "RemoveContainer" containerID="a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba" Mar 13 08:48:37 crc kubenswrapper[4876]: E0313 08:48:37.408477 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba\": container with ID starting with a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba not found: ID does not exist" containerID="a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.408551 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba"} err="failed to get container status \"a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba\": rpc error: code = NotFound desc = could not find container \"a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba\": container with ID starting with a75dac5b76a6adff9b7b62d5caef7bfa2017c416fc48720fdb314f7a988942ba not found: ID does not exist" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.408590 4876 scope.go:117] "RemoveContainer" containerID="60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161" Mar 13 08:48:37 crc kubenswrapper[4876]: E0313 08:48:37.409040 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161\": container with ID starting with 60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161 not found: ID does not exist" containerID="60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.409079 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161"} err="failed to get container status \"60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161\": rpc error: code = NotFound desc = could not find container \"60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161\": container with ID starting with 60c0da4914c499d9ee55e7896181cc0333f112989a29c1b53c557aa852a36161 not found: ID does not exist" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.409096 4876 scope.go:117] "RemoveContainer" containerID="8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f" Mar 13 08:48:37 crc kubenswrapper[4876]: E0313 08:48:37.409504 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f\": container with ID starting with 8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f not found: ID does not exist" containerID="8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f" Mar 13 08:48:37 crc kubenswrapper[4876]: I0313 08:48:37.409559 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f"} err="failed to get container status \"8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f\": rpc error: code = NotFound desc = could not find container \"8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f\": container with ID starting with 8fae2efd12f5a6adb0df5f0438c6ffc2422cf5ee5305e25cebc9da5451ae4a4f not found: ID does not exist" Mar 13 08:48:39 crc kubenswrapper[4876]: I0313 08:48:39.058972 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" path="/var/lib/kubelet/pods/7378eb49-c592-449a-82d3-07f9e7469b78/volumes" Mar 13 08:48:54 crc kubenswrapper[4876]: I0313 08:48:54.607032 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:48:54 crc kubenswrapper[4876]: I0313 08:48:54.607528 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:48:54 crc kubenswrapper[4876]: I0313 08:48:54.607571 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:48:54 crc kubenswrapper[4876]: I0313 08:48:54.608216 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f102aa15a16bff8287b68350bdd63c56ec0a2889615f66559f3f6993ac5c24a"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:48:54 crc kubenswrapper[4876]: I0313 08:48:54.608278 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://5f102aa15a16bff8287b68350bdd63c56ec0a2889615f66559f3f6993ac5c24a" gracePeriod=600 Mar 13 08:48:55 crc kubenswrapper[4876]: I0313 08:48:55.513147 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="5f102aa15a16bff8287b68350bdd63c56ec0a2889615f66559f3f6993ac5c24a" exitCode=0 Mar 13 08:48:55 crc kubenswrapper[4876]: I0313 08:48:55.513224 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"5f102aa15a16bff8287b68350bdd63c56ec0a2889615f66559f3f6993ac5c24a"} Mar 13 08:48:55 crc kubenswrapper[4876]: I0313 08:48:55.513673 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerStarted","Data":"4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1"} Mar 13 08:48:55 crc kubenswrapper[4876]: I0313 08:48:55.513699 4876 scope.go:117] "RemoveContainer" containerID="e665dd358f25cbb423fcdfffe22862538512753be98b671be4b7c22505364a55" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.141201 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556530-f84kb"] Mar 13 08:50:00 crc kubenswrapper[4876]: E0313 08:50:00.142268 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" containerName="extract-utilities" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.142284 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" containerName="extract-utilities" Mar 13 08:50:00 crc kubenswrapper[4876]: E0313 08:50:00.142310 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" containerName="extract-content" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.142317 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" containerName="extract-content" Mar 13 08:50:00 crc kubenswrapper[4876]: E0313 08:50:00.142329 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" containerName="registry-server" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.142337 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" containerName="registry-server" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.142600 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7378eb49-c592-449a-82d3-07f9e7469b78" containerName="registry-server" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.143347 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556530-f84kb" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.147206 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.147297 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.147680 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.150127 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556530-f84kb"] Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.258634 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfsjf\" (UniqueName: \"kubernetes.io/projected/c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21-kube-api-access-gfsjf\") pod \"auto-csr-approver-29556530-f84kb\" (UID: \"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21\") " pod="openshift-infra/auto-csr-approver-29556530-f84kb" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.360403 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfsjf\" (UniqueName: \"kubernetes.io/projected/c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21-kube-api-access-gfsjf\") pod \"auto-csr-approver-29556530-f84kb\" (UID: \"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21\") " pod="openshift-infra/auto-csr-approver-29556530-f84kb" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.384149 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfsjf\" (UniqueName: \"kubernetes.io/projected/c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21-kube-api-access-gfsjf\") pod \"auto-csr-approver-29556530-f84kb\" (UID: \"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21\") " pod="openshift-infra/auto-csr-approver-29556530-f84kb" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.466691 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556530-f84kb" Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.918191 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556530-f84kb"] Mar 13 08:50:00 crc kubenswrapper[4876]: I0313 08:50:00.933498 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 08:50:01 crc kubenswrapper[4876]: I0313 08:50:01.308294 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556530-f84kb" event={"ID":"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21","Type":"ContainerStarted","Data":"460af7dd361d8c762b857ae98faf2a5f983961a476536e7cd71668b8a5e66199"} Mar 13 08:50:02 crc kubenswrapper[4876]: I0313 08:50:02.317247 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556530-f84kb" event={"ID":"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21","Type":"ContainerStarted","Data":"633d459e1c42dbc9a2771ff19a8171cb34323b01e0c9546307a8ad933a31be01"} Mar 13 08:50:02 crc kubenswrapper[4876]: I0313 08:50:02.330657 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556530-f84kb" podStartSLOduration=1.2689858809999999 podStartE2EDuration="2.330637044s" podCreationTimestamp="2026-03-13 08:50:00 +0000 UTC" firstStartedPulling="2026-03-13 08:50:00.933050384 +0000 UTC m=+4260.603829406" lastFinishedPulling="2026-03-13 08:50:01.994701577 +0000 UTC m=+4261.665480569" observedRunningTime="2026-03-13 08:50:02.327657423 +0000 UTC m=+4261.998436405" watchObservedRunningTime="2026-03-13 08:50:02.330637044 +0000 UTC m=+4262.001416046" Mar 13 08:50:03 crc kubenswrapper[4876]: I0313 08:50:03.327693 4876 generic.go:334] "Generic (PLEG): container finished" podID="c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21" containerID="633d459e1c42dbc9a2771ff19a8171cb34323b01e0c9546307a8ad933a31be01" exitCode=0 Mar 13 08:50:03 crc kubenswrapper[4876]: I0313 08:50:03.328103 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556530-f84kb" event={"ID":"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21","Type":"ContainerDied","Data":"633d459e1c42dbc9a2771ff19a8171cb34323b01e0c9546307a8ad933a31be01"} Mar 13 08:50:04 crc kubenswrapper[4876]: I0313 08:50:04.744952 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556530-f84kb" Mar 13 08:50:04 crc kubenswrapper[4876]: I0313 08:50:04.853667 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfsjf\" (UniqueName: \"kubernetes.io/projected/c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21-kube-api-access-gfsjf\") pod \"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21\" (UID: \"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21\") " Mar 13 08:50:04 crc kubenswrapper[4876]: I0313 08:50:04.863758 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21-kube-api-access-gfsjf" (OuterVolumeSpecName: "kube-api-access-gfsjf") pod "c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21" (UID: "c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21"). InnerVolumeSpecName "kube-api-access-gfsjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:50:04 crc kubenswrapper[4876]: I0313 08:50:04.956569 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfsjf\" (UniqueName: \"kubernetes.io/projected/c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21-kube-api-access-gfsjf\") on node \"crc\" DevicePath \"\"" Mar 13 08:50:05 crc kubenswrapper[4876]: I0313 08:50:05.351954 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556530-f84kb" Mar 13 08:50:05 crc kubenswrapper[4876]: I0313 08:50:05.351914 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556530-f84kb" event={"ID":"c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21","Type":"ContainerDied","Data":"460af7dd361d8c762b857ae98faf2a5f983961a476536e7cd71668b8a5e66199"} Mar 13 08:50:05 crc kubenswrapper[4876]: I0313 08:50:05.352200 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="460af7dd361d8c762b857ae98faf2a5f983961a476536e7cd71668b8a5e66199" Mar 13 08:50:05 crc kubenswrapper[4876]: I0313 08:50:05.418666 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556524-gqckd"] Mar 13 08:50:05 crc kubenswrapper[4876]: I0313 08:50:05.429344 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556524-gqckd"] Mar 13 08:50:07 crc kubenswrapper[4876]: I0313 08:50:07.048050 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="773a0308-bb40-4075-b62e-2a3abbd04de9" path="/var/lib/kubelet/pods/773a0308-bb40-4075-b62e-2a3abbd04de9/volumes" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.190809 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2pwzc"] Mar 13 08:50:24 crc kubenswrapper[4876]: E0313 08:50:24.192144 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21" containerName="oc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.192162 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21" containerName="oc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.192423 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4a2dae7-71ef-48b5-9b84-b2ee0fa57b21" containerName="oc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.194332 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.211372 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2pwzc"] Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.322429 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65rmp\" (UniqueName: \"kubernetes.io/projected/59120207-5c57-4c12-bfc6-4a317e63fddb-kube-api-access-65rmp\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.322538 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-utilities\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.322569 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-catalog-content\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.425799 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-catalog-content\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.426078 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65rmp\" (UniqueName: \"kubernetes.io/projected/59120207-5c57-4c12-bfc6-4a317e63fddb-kube-api-access-65rmp\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.426317 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-utilities\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.426687 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-catalog-content\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.426795 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-utilities\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.450648 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65rmp\" (UniqueName: \"kubernetes.io/projected/59120207-5c57-4c12-bfc6-4a317e63fddb-kube-api-access-65rmp\") pod \"redhat-operators-2pwzc\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.524354 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.586326 4876 generic.go:334] "Generic (PLEG): container finished" podID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerID="f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35" exitCode=0 Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.586377 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wgfhk/must-gather-th6r7" event={"ID":"cc42ec18-ef73-4e58-831e-e6a2ead92f53","Type":"ContainerDied","Data":"f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35"} Mar 13 08:50:24 crc kubenswrapper[4876]: I0313 08:50:24.587074 4876 scope.go:117] "RemoveContainer" containerID="f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35" Mar 13 08:50:25 crc kubenswrapper[4876]: I0313 08:50:25.060021 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2pwzc"] Mar 13 08:50:25 crc kubenswrapper[4876]: I0313 08:50:25.544513 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wgfhk_must-gather-th6r7_cc42ec18-ef73-4e58-831e-e6a2ead92f53/gather/0.log" Mar 13 08:50:25 crc kubenswrapper[4876]: I0313 08:50:25.597288 4876 generic.go:334] "Generic (PLEG): container finished" podID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerID="0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3" exitCode=0 Mar 13 08:50:25 crc kubenswrapper[4876]: I0313 08:50:25.597358 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2pwzc" event={"ID":"59120207-5c57-4c12-bfc6-4a317e63fddb","Type":"ContainerDied","Data":"0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3"} Mar 13 08:50:25 crc kubenswrapper[4876]: I0313 08:50:25.597402 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2pwzc" event={"ID":"59120207-5c57-4c12-bfc6-4a317e63fddb","Type":"ContainerStarted","Data":"0e09c9534496a1c2a5a3fb600a12e7b57999cc6f65f3d3e1627244246d4de180"} Mar 13 08:50:26 crc kubenswrapper[4876]: I0313 08:50:26.610868 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2pwzc" event={"ID":"59120207-5c57-4c12-bfc6-4a317e63fddb","Type":"ContainerStarted","Data":"5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b"} Mar 13 08:50:27 crc kubenswrapper[4876]: I0313 08:50:27.630760 4876 generic.go:334] "Generic (PLEG): container finished" podID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerID="5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b" exitCode=0 Mar 13 08:50:27 crc kubenswrapper[4876]: I0313 08:50:27.630991 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2pwzc" event={"ID":"59120207-5c57-4c12-bfc6-4a317e63fddb","Type":"ContainerDied","Data":"5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b"} Mar 13 08:50:28 crc kubenswrapper[4876]: I0313 08:50:28.675650 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2pwzc" event={"ID":"59120207-5c57-4c12-bfc6-4a317e63fddb","Type":"ContainerStarted","Data":"ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d"} Mar 13 08:50:28 crc kubenswrapper[4876]: I0313 08:50:28.710770 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2pwzc" podStartSLOduration=2.144635086 podStartE2EDuration="4.710747653s" podCreationTimestamp="2026-03-13 08:50:24 +0000 UTC" firstStartedPulling="2026-03-13 08:50:25.601088261 +0000 UTC m=+4285.271867263" lastFinishedPulling="2026-03-13 08:50:28.167200828 +0000 UTC m=+4287.837979830" observedRunningTime="2026-03-13 08:50:28.701530751 +0000 UTC m=+4288.372309733" watchObservedRunningTime="2026-03-13 08:50:28.710747653 +0000 UTC m=+4288.381526635" Mar 13 08:50:30 crc kubenswrapper[4876]: I0313 08:50:30.234875 4876 scope.go:117] "RemoveContainer" containerID="be7bfa08a56eeacbc1bece6fc207ee653035be7632a9032ec670a6d7a022a0cb" Mar 13 08:50:34 crc kubenswrapper[4876]: I0313 08:50:34.526343 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:34 crc kubenswrapper[4876]: I0313 08:50:34.526954 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:35 crc kubenswrapper[4876]: I0313 08:50:35.582177 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2pwzc" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="registry-server" probeResult="failure" output=< Mar 13 08:50:35 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Mar 13 08:50:35 crc kubenswrapper[4876]: > Mar 13 08:50:35 crc kubenswrapper[4876]: I0313 08:50:35.987252 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wgfhk/must-gather-th6r7"] Mar 13 08:50:35 crc kubenswrapper[4876]: I0313 08:50:35.987644 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wgfhk/must-gather-th6r7" podUID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerName="copy" containerID="cri-o://eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f" gracePeriod=2 Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.003790 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wgfhk/must-gather-th6r7"] Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.481620 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wgfhk_must-gather-th6r7_cc42ec18-ef73-4e58-831e-e6a2ead92f53/copy/0.log" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.482401 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.524587 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfzz6\" (UniqueName: \"kubernetes.io/projected/cc42ec18-ef73-4e58-831e-e6a2ead92f53-kube-api-access-zfzz6\") pod \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\" (UID: \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\") " Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.524748 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cc42ec18-ef73-4e58-831e-e6a2ead92f53-must-gather-output\") pod \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\" (UID: \"cc42ec18-ef73-4e58-831e-e6a2ead92f53\") " Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.545519 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc42ec18-ef73-4e58-831e-e6a2ead92f53-kube-api-access-zfzz6" (OuterVolumeSpecName: "kube-api-access-zfzz6") pod "cc42ec18-ef73-4e58-831e-e6a2ead92f53" (UID: "cc42ec18-ef73-4e58-831e-e6a2ead92f53"). InnerVolumeSpecName "kube-api-access-zfzz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.628646 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfzz6\" (UniqueName: \"kubernetes.io/projected/cc42ec18-ef73-4e58-831e-e6a2ead92f53-kube-api-access-zfzz6\") on node \"crc\" DevicePath \"\"" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.703949 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc42ec18-ef73-4e58-831e-e6a2ead92f53-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cc42ec18-ef73-4e58-831e-e6a2ead92f53" (UID: "cc42ec18-ef73-4e58-831e-e6a2ead92f53"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.730543 4876 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cc42ec18-ef73-4e58-831e-e6a2ead92f53-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.848358 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wgfhk_must-gather-th6r7_cc42ec18-ef73-4e58-831e-e6a2ead92f53/copy/0.log" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.849123 4876 generic.go:334] "Generic (PLEG): container finished" podID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerID="eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f" exitCode=143 Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.849229 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wgfhk/must-gather-th6r7" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.849281 4876 scope.go:117] "RemoveContainer" containerID="eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.881076 4876 scope.go:117] "RemoveContainer" containerID="f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.996877 4876 scope.go:117] "RemoveContainer" containerID="eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f" Mar 13 08:50:36 crc kubenswrapper[4876]: E0313 08:50:36.997711 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f\": container with ID starting with eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f not found: ID does not exist" containerID="eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.997793 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f"} err="failed to get container status \"eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f\": rpc error: code = NotFound desc = could not find container \"eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f\": container with ID starting with eddd50d7457663f63d2c0dbb77e53bc8199255890e44d873dadf00e0dda1aa6f not found: ID does not exist" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.997842 4876 scope.go:117] "RemoveContainer" containerID="f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35" Mar 13 08:50:36 crc kubenswrapper[4876]: E0313 08:50:36.998536 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35\": container with ID starting with f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35 not found: ID does not exist" containerID="f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35" Mar 13 08:50:36 crc kubenswrapper[4876]: I0313 08:50:36.998597 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35"} err="failed to get container status \"f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35\": rpc error: code = NotFound desc = could not find container \"f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35\": container with ID starting with f21b7d8683b257ba216e200f7e45cf53f98d0f7d8dd76af9a3111bdfa3e1fb35 not found: ID does not exist" Mar 13 08:50:37 crc kubenswrapper[4876]: I0313 08:50:37.047309 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" path="/var/lib/kubelet/pods/cc42ec18-ef73-4e58-831e-e6a2ead92f53/volumes" Mar 13 08:50:44 crc kubenswrapper[4876]: I0313 08:50:44.581743 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:44 crc kubenswrapper[4876]: I0313 08:50:44.634337 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:44 crc kubenswrapper[4876]: I0313 08:50:44.818361 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2pwzc"] Mar 13 08:50:45 crc kubenswrapper[4876]: I0313 08:50:45.938746 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2pwzc" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="registry-server" containerID="cri-o://ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d" gracePeriod=2 Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.410681 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.465836 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-catalog-content\") pod \"59120207-5c57-4c12-bfc6-4a317e63fddb\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.466063 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-utilities\") pod \"59120207-5c57-4c12-bfc6-4a317e63fddb\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.466338 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65rmp\" (UniqueName: \"kubernetes.io/projected/59120207-5c57-4c12-bfc6-4a317e63fddb-kube-api-access-65rmp\") pod \"59120207-5c57-4c12-bfc6-4a317e63fddb\" (UID: \"59120207-5c57-4c12-bfc6-4a317e63fddb\") " Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.469535 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-utilities" (OuterVolumeSpecName: "utilities") pod "59120207-5c57-4c12-bfc6-4a317e63fddb" (UID: "59120207-5c57-4c12-bfc6-4a317e63fddb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.477326 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59120207-5c57-4c12-bfc6-4a317e63fddb-kube-api-access-65rmp" (OuterVolumeSpecName: "kube-api-access-65rmp") pod "59120207-5c57-4c12-bfc6-4a317e63fddb" (UID: "59120207-5c57-4c12-bfc6-4a317e63fddb"). InnerVolumeSpecName "kube-api-access-65rmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.568905 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65rmp\" (UniqueName: \"kubernetes.io/projected/59120207-5c57-4c12-bfc6-4a317e63fddb-kube-api-access-65rmp\") on node \"crc\" DevicePath \"\"" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.568935 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.640591 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59120207-5c57-4c12-bfc6-4a317e63fddb" (UID: "59120207-5c57-4c12-bfc6-4a317e63fddb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.672795 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59120207-5c57-4c12-bfc6-4a317e63fddb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.956529 4876 generic.go:334] "Generic (PLEG): container finished" podID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerID="ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d" exitCode=0 Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.956605 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2pwzc" event={"ID":"59120207-5c57-4c12-bfc6-4a317e63fddb","Type":"ContainerDied","Data":"ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d"} Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.956654 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2pwzc" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.956699 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2pwzc" event={"ID":"59120207-5c57-4c12-bfc6-4a317e63fddb","Type":"ContainerDied","Data":"0e09c9534496a1c2a5a3fb600a12e7b57999cc6f65f3d3e1627244246d4de180"} Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.956730 4876 scope.go:117] "RemoveContainer" containerID="ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d" Mar 13 08:50:46 crc kubenswrapper[4876]: I0313 08:50:46.986043 4876 scope.go:117] "RemoveContainer" containerID="5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b" Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.001462 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2pwzc"] Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.010084 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2pwzc"] Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.027093 4876 scope.go:117] "RemoveContainer" containerID="0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3" Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.047105 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" path="/var/lib/kubelet/pods/59120207-5c57-4c12-bfc6-4a317e63fddb/volumes" Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.065786 4876 scope.go:117] "RemoveContainer" containerID="ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d" Mar 13 08:50:47 crc kubenswrapper[4876]: E0313 08:50:47.066404 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d\": container with ID starting with ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d not found: ID does not exist" containerID="ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d" Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.066466 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d"} err="failed to get container status \"ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d\": rpc error: code = NotFound desc = could not find container \"ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d\": container with ID starting with ba569b07b2f9bccc652840255f2371f9acb1b9538f3455ff59dab74c3194447d not found: ID does not exist" Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.066493 4876 scope.go:117] "RemoveContainer" containerID="5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b" Mar 13 08:50:47 crc kubenswrapper[4876]: E0313 08:50:47.066847 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b\": container with ID starting with 5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b not found: ID does not exist" containerID="5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b" Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.066887 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b"} err="failed to get container status \"5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b\": rpc error: code = NotFound desc = could not find container \"5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b\": container with ID starting with 5057e019a4486fe6a368c81d89ff4fc83974634e736c2dbefa1116964c28d53b not found: ID does not exist" Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.066914 4876 scope.go:117] "RemoveContainer" containerID="0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3" Mar 13 08:50:47 crc kubenswrapper[4876]: E0313 08:50:47.067261 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3\": container with ID starting with 0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3 not found: ID does not exist" containerID="0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3" Mar 13 08:50:47 crc kubenswrapper[4876]: I0313 08:50:47.067287 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3"} err="failed to get container status \"0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3\": rpc error: code = NotFound desc = could not find container \"0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3\": container with ID starting with 0f0e439237b1a9cb115bda5fe0c15489da7cd0a6534a4c794fdad88e112f6ea3 not found: ID does not exist" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.904491 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dhztg"] Mar 13 08:51:16 crc kubenswrapper[4876]: E0313 08:51:16.905619 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="extract-utilities" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.905650 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="extract-utilities" Mar 13 08:51:16 crc kubenswrapper[4876]: E0313 08:51:16.905681 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerName="copy" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.905693 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerName="copy" Mar 13 08:51:16 crc kubenswrapper[4876]: E0313 08:51:16.905730 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="extract-content" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.905744 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="extract-content" Mar 13 08:51:16 crc kubenswrapper[4876]: E0313 08:51:16.905767 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerName="gather" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.905780 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerName="gather" Mar 13 08:51:16 crc kubenswrapper[4876]: E0313 08:51:16.905797 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="registry-server" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.905809 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="registry-server" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.906230 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerName="gather" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.906329 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="59120207-5c57-4c12-bfc6-4a317e63fddb" containerName="registry-server" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.906358 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc42ec18-ef73-4e58-831e-e6a2ead92f53" containerName="copy" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.909019 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.922943 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dhztg"] Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.969089 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e891f526-4cec-4e10-89e9-158015a1d7bd-catalog-content\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.969276 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e891f526-4cec-4e10-89e9-158015a1d7bd-utilities\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:16 crc kubenswrapper[4876]: I0313 08:51:16.969345 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rgwg\" (UniqueName: \"kubernetes.io/projected/e891f526-4cec-4e10-89e9-158015a1d7bd-kube-api-access-9rgwg\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:17 crc kubenswrapper[4876]: I0313 08:51:17.071792 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e891f526-4cec-4e10-89e9-158015a1d7bd-utilities\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:17 crc kubenswrapper[4876]: I0313 08:51:17.071906 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rgwg\" (UniqueName: \"kubernetes.io/projected/e891f526-4cec-4e10-89e9-158015a1d7bd-kube-api-access-9rgwg\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:17 crc kubenswrapper[4876]: I0313 08:51:17.071960 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e891f526-4cec-4e10-89e9-158015a1d7bd-catalog-content\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:17 crc kubenswrapper[4876]: I0313 08:51:17.072412 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e891f526-4cec-4e10-89e9-158015a1d7bd-utilities\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:17 crc kubenswrapper[4876]: I0313 08:51:17.072649 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e891f526-4cec-4e10-89e9-158015a1d7bd-catalog-content\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:17 crc kubenswrapper[4876]: I0313 08:51:17.096957 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rgwg\" (UniqueName: \"kubernetes.io/projected/e891f526-4cec-4e10-89e9-158015a1d7bd-kube-api-access-9rgwg\") pod \"community-operators-dhztg\" (UID: \"e891f526-4cec-4e10-89e9-158015a1d7bd\") " pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:17 crc kubenswrapper[4876]: I0313 08:51:17.237041 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:17 crc kubenswrapper[4876]: I0313 08:51:17.809114 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dhztg"] Mar 13 08:51:18 crc kubenswrapper[4876]: I0313 08:51:18.300105 4876 generic.go:334] "Generic (PLEG): container finished" podID="e891f526-4cec-4e10-89e9-158015a1d7bd" containerID="cfd778fff71b094c8e8f551e75cddad97c851dba7f493c3a9b3d23cd623bfa71" exitCode=0 Mar 13 08:51:18 crc kubenswrapper[4876]: I0313 08:51:18.300148 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhztg" event={"ID":"e891f526-4cec-4e10-89e9-158015a1d7bd","Type":"ContainerDied","Data":"cfd778fff71b094c8e8f551e75cddad97c851dba7f493c3a9b3d23cd623bfa71"} Mar 13 08:51:18 crc kubenswrapper[4876]: I0313 08:51:18.300172 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhztg" event={"ID":"e891f526-4cec-4e10-89e9-158015a1d7bd","Type":"ContainerStarted","Data":"c08670dd6bb6243f740692a6ad0366b3db0b0c264fe65e7b1417c6721567ffbd"} Mar 13 08:51:23 crc kubenswrapper[4876]: I0313 08:51:23.348414 4876 generic.go:334] "Generic (PLEG): container finished" podID="e891f526-4cec-4e10-89e9-158015a1d7bd" containerID="e124e3242eac2fc8c2a42ea5a6cb0bcfc4e060ccac18e6302e24f0de52d8c78c" exitCode=0 Mar 13 08:51:23 crc kubenswrapper[4876]: I0313 08:51:23.348482 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhztg" event={"ID":"e891f526-4cec-4e10-89e9-158015a1d7bd","Type":"ContainerDied","Data":"e124e3242eac2fc8c2a42ea5a6cb0bcfc4e060ccac18e6302e24f0de52d8c78c"} Mar 13 08:51:24 crc kubenswrapper[4876]: I0313 08:51:24.361260 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhztg" event={"ID":"e891f526-4cec-4e10-89e9-158015a1d7bd","Type":"ContainerStarted","Data":"fdb841dbf1ca05fe2326f3c985c5809e4bac2f7ee42b109018f67a533d9bdaf9"} Mar 13 08:51:24 crc kubenswrapper[4876]: I0313 08:51:24.406094 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dhztg" podStartSLOduration=2.6965905980000002 podStartE2EDuration="8.406068179s" podCreationTimestamp="2026-03-13 08:51:16 +0000 UTC" firstStartedPulling="2026-03-13 08:51:18.305773669 +0000 UTC m=+4337.976552681" lastFinishedPulling="2026-03-13 08:51:24.01525128 +0000 UTC m=+4343.686030262" observedRunningTime="2026-03-13 08:51:24.394710027 +0000 UTC m=+4344.065489039" watchObservedRunningTime="2026-03-13 08:51:24.406068179 +0000 UTC m=+4344.076847161" Mar 13 08:51:24 crc kubenswrapper[4876]: I0313 08:51:24.608268 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:51:24 crc kubenswrapper[4876]: I0313 08:51:24.608336 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:51:27 crc kubenswrapper[4876]: I0313 08:51:27.237430 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:27 crc kubenswrapper[4876]: I0313 08:51:27.239367 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:27 crc kubenswrapper[4876]: I0313 08:51:27.313591 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:37 crc kubenswrapper[4876]: I0313 08:51:37.337500 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dhztg" Mar 13 08:51:37 crc kubenswrapper[4876]: I0313 08:51:37.438195 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dhztg"] Mar 13 08:51:37 crc kubenswrapper[4876]: I0313 08:51:37.499330 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxfhc"] Mar 13 08:51:37 crc kubenswrapper[4876]: I0313 08:51:37.499655 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pxfhc" podUID="add82187-d01b-4902-8f12-ced45327d1f9" containerName="registry-server" containerID="cri-o://18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e" gracePeriod=2 Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.026229 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxfhc" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.093214 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-catalog-content\") pod \"add82187-d01b-4902-8f12-ced45327d1f9\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.093313 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-utilities\") pod \"add82187-d01b-4902-8f12-ced45327d1f9\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.093423 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxv4m\" (UniqueName: \"kubernetes.io/projected/add82187-d01b-4902-8f12-ced45327d1f9-kube-api-access-pxv4m\") pod \"add82187-d01b-4902-8f12-ced45327d1f9\" (UID: \"add82187-d01b-4902-8f12-ced45327d1f9\") " Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.099294 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-utilities" (OuterVolumeSpecName: "utilities") pod "add82187-d01b-4902-8f12-ced45327d1f9" (UID: "add82187-d01b-4902-8f12-ced45327d1f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.125052 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/add82187-d01b-4902-8f12-ced45327d1f9-kube-api-access-pxv4m" (OuterVolumeSpecName: "kube-api-access-pxv4m") pod "add82187-d01b-4902-8f12-ced45327d1f9" (UID: "add82187-d01b-4902-8f12-ced45327d1f9"). InnerVolumeSpecName "kube-api-access-pxv4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.173726 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "add82187-d01b-4902-8f12-ced45327d1f9" (UID: "add82187-d01b-4902-8f12-ced45327d1f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.200528 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.200572 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/add82187-d01b-4902-8f12-ced45327d1f9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.200587 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxv4m\" (UniqueName: \"kubernetes.io/projected/add82187-d01b-4902-8f12-ced45327d1f9-kube-api-access-pxv4m\") on node \"crc\" DevicePath \"\"" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.523542 4876 generic.go:334] "Generic (PLEG): container finished" podID="add82187-d01b-4902-8f12-ced45327d1f9" containerID="18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e" exitCode=0 Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.523595 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxfhc" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.523598 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxfhc" event={"ID":"add82187-d01b-4902-8f12-ced45327d1f9","Type":"ContainerDied","Data":"18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e"} Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.523746 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxfhc" event={"ID":"add82187-d01b-4902-8f12-ced45327d1f9","Type":"ContainerDied","Data":"decc1a185f5ebc48cf055b8bda5f932c702c82617187fadb1cac20e501d8d29e"} Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.523770 4876 scope.go:117] "RemoveContainer" containerID="18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.545011 4876 scope.go:117] "RemoveContainer" containerID="c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.557942 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxfhc"] Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.573625 4876 scope.go:117] "RemoveContainer" containerID="2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.573887 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pxfhc"] Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.618487 4876 scope.go:117] "RemoveContainer" containerID="18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e" Mar 13 08:51:38 crc kubenswrapper[4876]: E0313 08:51:38.618989 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e\": container with ID starting with 18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e not found: ID does not exist" containerID="18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.619039 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e"} err="failed to get container status \"18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e\": rpc error: code = NotFound desc = could not find container \"18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e\": container with ID starting with 18785c73a2b6ffc7f7e182c9e0b0e7e9f107d62bc796396ab69f171297b6792e not found: ID does not exist" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.619070 4876 scope.go:117] "RemoveContainer" containerID="c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d" Mar 13 08:51:38 crc kubenswrapper[4876]: E0313 08:51:38.619585 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d\": container with ID starting with c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d not found: ID does not exist" containerID="c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.619626 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d"} err="failed to get container status \"c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d\": rpc error: code = NotFound desc = could not find container \"c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d\": container with ID starting with c92eecaa6a6a8f4b5e864da036c24aff47c6b24beb276ccc160a63589154095d not found: ID does not exist" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.619655 4876 scope.go:117] "RemoveContainer" containerID="2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511" Mar 13 08:51:38 crc kubenswrapper[4876]: E0313 08:51:38.619990 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511\": container with ID starting with 2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511 not found: ID does not exist" containerID="2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511" Mar 13 08:51:38 crc kubenswrapper[4876]: I0313 08:51:38.620043 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511"} err="failed to get container status \"2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511\": rpc error: code = NotFound desc = could not find container \"2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511\": container with ID starting with 2687271a4a2dba9500396b861b14a9dbc1447ba76b4e413daee5c99e7678f511 not found: ID does not exist" Mar 13 08:51:39 crc kubenswrapper[4876]: I0313 08:51:39.049084 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="add82187-d01b-4902-8f12-ced45327d1f9" path="/var/lib/kubelet/pods/add82187-d01b-4902-8f12-ced45327d1f9/volumes" Mar 13 08:51:54 crc kubenswrapper[4876]: I0313 08:51:54.607116 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:51:54 crc kubenswrapper[4876]: I0313 08:51:54.607975 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.169905 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556532-jcgz9"] Mar 13 08:52:00 crc kubenswrapper[4876]: E0313 08:52:00.171162 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="add82187-d01b-4902-8f12-ced45327d1f9" containerName="extract-content" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.171184 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="add82187-d01b-4902-8f12-ced45327d1f9" containerName="extract-content" Mar 13 08:52:00 crc kubenswrapper[4876]: E0313 08:52:00.171269 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="add82187-d01b-4902-8f12-ced45327d1f9" containerName="extract-utilities" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.171283 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="add82187-d01b-4902-8f12-ced45327d1f9" containerName="extract-utilities" Mar 13 08:52:00 crc kubenswrapper[4876]: E0313 08:52:00.171306 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="add82187-d01b-4902-8f12-ced45327d1f9" containerName="registry-server" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.171318 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="add82187-d01b-4902-8f12-ced45327d1f9" containerName="registry-server" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.171620 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="add82187-d01b-4902-8f12-ced45327d1f9" containerName="registry-server" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.172525 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556532-jcgz9" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.182104 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556532-jcgz9"] Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.190267 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.190304 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.190337 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.275126 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtn92\" (UniqueName: \"kubernetes.io/projected/181746a2-077b-4099-8f0a-81df37591809-kube-api-access-wtn92\") pod \"auto-csr-approver-29556532-jcgz9\" (UID: \"181746a2-077b-4099-8f0a-81df37591809\") " pod="openshift-infra/auto-csr-approver-29556532-jcgz9" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.376691 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtn92\" (UniqueName: \"kubernetes.io/projected/181746a2-077b-4099-8f0a-81df37591809-kube-api-access-wtn92\") pod \"auto-csr-approver-29556532-jcgz9\" (UID: \"181746a2-077b-4099-8f0a-81df37591809\") " pod="openshift-infra/auto-csr-approver-29556532-jcgz9" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.396539 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtn92\" (UniqueName: \"kubernetes.io/projected/181746a2-077b-4099-8f0a-81df37591809-kube-api-access-wtn92\") pod \"auto-csr-approver-29556532-jcgz9\" (UID: \"181746a2-077b-4099-8f0a-81df37591809\") " pod="openshift-infra/auto-csr-approver-29556532-jcgz9" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.522727 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556532-jcgz9" Mar 13 08:52:00 crc kubenswrapper[4876]: I0313 08:52:00.971949 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556532-jcgz9"] Mar 13 08:52:01 crc kubenswrapper[4876]: I0313 08:52:01.801842 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556532-jcgz9" event={"ID":"181746a2-077b-4099-8f0a-81df37591809","Type":"ContainerStarted","Data":"b35ac9dec18a906fcad5f218f763453c8aa437d036fc8cc7babce67c4c11b6b3"} Mar 13 08:52:02 crc kubenswrapper[4876]: I0313 08:52:02.812299 4876 generic.go:334] "Generic (PLEG): container finished" podID="181746a2-077b-4099-8f0a-81df37591809" containerID="34774062e81345f5324f5ac4235e362329ba1f93ad8f7be7ba6228dc53c896eb" exitCode=0 Mar 13 08:52:02 crc kubenswrapper[4876]: I0313 08:52:02.812344 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556532-jcgz9" event={"ID":"181746a2-077b-4099-8f0a-81df37591809","Type":"ContainerDied","Data":"34774062e81345f5324f5ac4235e362329ba1f93ad8f7be7ba6228dc53c896eb"} Mar 13 08:52:04 crc kubenswrapper[4876]: I0313 08:52:04.197467 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556532-jcgz9" Mar 13 08:52:04 crc kubenswrapper[4876]: I0313 08:52:04.361126 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtn92\" (UniqueName: \"kubernetes.io/projected/181746a2-077b-4099-8f0a-81df37591809-kube-api-access-wtn92\") pod \"181746a2-077b-4099-8f0a-81df37591809\" (UID: \"181746a2-077b-4099-8f0a-81df37591809\") " Mar 13 08:52:04 crc kubenswrapper[4876]: I0313 08:52:04.369495 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/181746a2-077b-4099-8f0a-81df37591809-kube-api-access-wtn92" (OuterVolumeSpecName: "kube-api-access-wtn92") pod "181746a2-077b-4099-8f0a-81df37591809" (UID: "181746a2-077b-4099-8f0a-81df37591809"). InnerVolumeSpecName "kube-api-access-wtn92". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:52:04 crc kubenswrapper[4876]: I0313 08:52:04.464362 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtn92\" (UniqueName: \"kubernetes.io/projected/181746a2-077b-4099-8f0a-81df37591809-kube-api-access-wtn92\") on node \"crc\" DevicePath \"\"" Mar 13 08:52:04 crc kubenswrapper[4876]: I0313 08:52:04.831527 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556532-jcgz9" event={"ID":"181746a2-077b-4099-8f0a-81df37591809","Type":"ContainerDied","Data":"b35ac9dec18a906fcad5f218f763453c8aa437d036fc8cc7babce67c4c11b6b3"} Mar 13 08:52:04 crc kubenswrapper[4876]: I0313 08:52:04.832056 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b35ac9dec18a906fcad5f218f763453c8aa437d036fc8cc7babce67c4c11b6b3" Mar 13 08:52:04 crc kubenswrapper[4876]: I0313 08:52:04.831606 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556532-jcgz9" Mar 13 08:52:05 crc kubenswrapper[4876]: I0313 08:52:05.270736 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556526-cgtsw"] Mar 13 08:52:05 crc kubenswrapper[4876]: I0313 08:52:05.277756 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556526-cgtsw"] Mar 13 08:52:07 crc kubenswrapper[4876]: I0313 08:52:07.057450 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c" path="/var/lib/kubelet/pods/b7e8f18b-40b0-439d-ba4c-5bbd66fdba3c/volumes" Mar 13 08:52:24 crc kubenswrapper[4876]: I0313 08:52:24.607550 4876 patch_prober.go:28] interesting pod/machine-config-daemon-r9cl2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 08:52:24 crc kubenswrapper[4876]: I0313 08:52:24.608006 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 08:52:24 crc kubenswrapper[4876]: I0313 08:52:24.608051 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" Mar 13 08:52:24 crc kubenswrapper[4876]: I0313 08:52:24.608879 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1"} pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 08:52:24 crc kubenswrapper[4876]: I0313 08:52:24.608935 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" containerName="machine-config-daemon" containerID="cri-o://4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" gracePeriod=600 Mar 13 08:52:24 crc kubenswrapper[4876]: E0313 08:52:24.731572 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:52:25 crc kubenswrapper[4876]: I0313 08:52:25.049444 4876 generic.go:334] "Generic (PLEG): container finished" podID="0a6f71e5-2091-4386-b559-bba70bc45972" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" exitCode=0 Mar 13 08:52:25 crc kubenswrapper[4876]: I0313 08:52:25.049689 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" event={"ID":"0a6f71e5-2091-4386-b559-bba70bc45972","Type":"ContainerDied","Data":"4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1"} Mar 13 08:52:25 crc kubenswrapper[4876]: I0313 08:52:25.049718 4876 scope.go:117] "RemoveContainer" containerID="5f102aa15a16bff8287b68350bdd63c56ec0a2889615f66559f3f6993ac5c24a" Mar 13 08:52:25 crc kubenswrapper[4876]: I0313 08:52:25.050282 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:52:25 crc kubenswrapper[4876]: E0313 08:52:25.050651 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:52:30 crc kubenswrapper[4876]: I0313 08:52:30.405032 4876 scope.go:117] "RemoveContainer" containerID="4bb87886aeeb890284330a4a0a11684d110cd4d348a38ed6c25974f7a40e3f86" Mar 13 08:52:38 crc kubenswrapper[4876]: I0313 08:52:38.035885 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:52:38 crc kubenswrapper[4876]: E0313 08:52:38.036783 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:52:53 crc kubenswrapper[4876]: I0313 08:52:53.042616 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:52:53 crc kubenswrapper[4876]: E0313 08:52:53.043440 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:53:06 crc kubenswrapper[4876]: I0313 08:53:06.036370 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:53:06 crc kubenswrapper[4876]: E0313 08:53:06.037433 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:53:20 crc kubenswrapper[4876]: I0313 08:53:20.035891 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:53:20 crc kubenswrapper[4876]: E0313 08:53:20.037341 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:53:35 crc kubenswrapper[4876]: I0313 08:53:35.035976 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:53:35 crc kubenswrapper[4876]: E0313 08:53:35.037168 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:53:50 crc kubenswrapper[4876]: I0313 08:53:50.035715 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:53:50 crc kubenswrapper[4876]: E0313 08:53:50.036492 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.178903 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556534-lfxsk"] Mar 13 08:54:00 crc kubenswrapper[4876]: E0313 08:54:00.180287 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181746a2-077b-4099-8f0a-81df37591809" containerName="oc" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.180310 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="181746a2-077b-4099-8f0a-81df37591809" containerName="oc" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.180643 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="181746a2-077b-4099-8f0a-81df37591809" containerName="oc" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.181674 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556534-lfxsk" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.184002 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.184287 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.189702 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brx67" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.190628 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556534-lfxsk"] Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.281405 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsklr\" (UniqueName: \"kubernetes.io/projected/f5d82515-9011-4fb0-ab54-934846a9480e-kube-api-access-wsklr\") pod \"auto-csr-approver-29556534-lfxsk\" (UID: \"f5d82515-9011-4fb0-ab54-934846a9480e\") " pod="openshift-infra/auto-csr-approver-29556534-lfxsk" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.383900 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsklr\" (UniqueName: \"kubernetes.io/projected/f5d82515-9011-4fb0-ab54-934846a9480e-kube-api-access-wsklr\") pod \"auto-csr-approver-29556534-lfxsk\" (UID: \"f5d82515-9011-4fb0-ab54-934846a9480e\") " pod="openshift-infra/auto-csr-approver-29556534-lfxsk" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.402684 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsklr\" (UniqueName: \"kubernetes.io/projected/f5d82515-9011-4fb0-ab54-934846a9480e-kube-api-access-wsklr\") pod \"auto-csr-approver-29556534-lfxsk\" (UID: \"f5d82515-9011-4fb0-ab54-934846a9480e\") " pod="openshift-infra/auto-csr-approver-29556534-lfxsk" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.508794 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556534-lfxsk" Mar 13 08:54:00 crc kubenswrapper[4876]: I0313 08:54:00.974111 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556534-lfxsk"] Mar 13 08:54:01 crc kubenswrapper[4876]: I0313 08:54:01.043754 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:54:01 crc kubenswrapper[4876]: E0313 08:54:01.044001 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:54:01 crc kubenswrapper[4876]: I0313 08:54:01.132709 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556534-lfxsk" event={"ID":"f5d82515-9011-4fb0-ab54-934846a9480e","Type":"ContainerStarted","Data":"51803334ecfb17be0e7927f2a81e1d90e47a8a9b8ff507d147f2633154b1eba6"} Mar 13 08:54:03 crc kubenswrapper[4876]: I0313 08:54:03.153013 4876 generic.go:334] "Generic (PLEG): container finished" podID="f5d82515-9011-4fb0-ab54-934846a9480e" containerID="cc737f3c2e1d611c63cb48bd1fb685585f09069723219c6a9af41f674744bb0d" exitCode=0 Mar 13 08:54:03 crc kubenswrapper[4876]: I0313 08:54:03.153088 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556534-lfxsk" event={"ID":"f5d82515-9011-4fb0-ab54-934846a9480e","Type":"ContainerDied","Data":"cc737f3c2e1d611c63cb48bd1fb685585f09069723219c6a9af41f674744bb0d"} Mar 13 08:54:04 crc kubenswrapper[4876]: I0313 08:54:04.687989 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556534-lfxsk" Mar 13 08:54:04 crc kubenswrapper[4876]: I0313 08:54:04.878218 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsklr\" (UniqueName: \"kubernetes.io/projected/f5d82515-9011-4fb0-ab54-934846a9480e-kube-api-access-wsklr\") pod \"f5d82515-9011-4fb0-ab54-934846a9480e\" (UID: \"f5d82515-9011-4fb0-ab54-934846a9480e\") " Mar 13 08:54:04 crc kubenswrapper[4876]: I0313 08:54:04.883466 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d82515-9011-4fb0-ab54-934846a9480e-kube-api-access-wsklr" (OuterVolumeSpecName: "kube-api-access-wsklr") pod "f5d82515-9011-4fb0-ab54-934846a9480e" (UID: "f5d82515-9011-4fb0-ab54-934846a9480e"). InnerVolumeSpecName "kube-api-access-wsklr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 08:54:04 crc kubenswrapper[4876]: I0313 08:54:04.982541 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsklr\" (UniqueName: \"kubernetes.io/projected/f5d82515-9011-4fb0-ab54-934846a9480e-kube-api-access-wsklr\") on node \"crc\" DevicePath \"\"" Mar 13 08:54:05 crc kubenswrapper[4876]: I0313 08:54:05.180727 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556534-lfxsk" event={"ID":"f5d82515-9011-4fb0-ab54-934846a9480e","Type":"ContainerDied","Data":"51803334ecfb17be0e7927f2a81e1d90e47a8a9b8ff507d147f2633154b1eba6"} Mar 13 08:54:05 crc kubenswrapper[4876]: I0313 08:54:05.181082 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51803334ecfb17be0e7927f2a81e1d90e47a8a9b8ff507d147f2633154b1eba6" Mar 13 08:54:05 crc kubenswrapper[4876]: I0313 08:54:05.180860 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556534-lfxsk" Mar 13 08:54:05 crc kubenswrapper[4876]: I0313 08:54:05.764373 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556528-6rz8d"] Mar 13 08:54:05 crc kubenswrapper[4876]: I0313 08:54:05.774180 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556528-6rz8d"] Mar 13 08:54:07 crc kubenswrapper[4876]: I0313 08:54:07.047721 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1acea7c3-7bed-454e-bfc5-90ea76fd4a51" path="/var/lib/kubelet/pods/1acea7c3-7bed-454e-bfc5-90ea76fd4a51/volumes" Mar 13 08:54:15 crc kubenswrapper[4876]: I0313 08:54:15.035167 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:54:15 crc kubenswrapper[4876]: E0313 08:54:15.036048 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:54:28 crc kubenswrapper[4876]: I0313 08:54:28.037826 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:54:28 crc kubenswrapper[4876]: E0313 08:54:28.038668 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" Mar 13 08:54:30 crc kubenswrapper[4876]: I0313 08:54:30.519751 4876 scope.go:117] "RemoveContainer" containerID="0af5c7cea6d8d1cdc7d3131732b0fc2ee1f360021c86c4cec9642b2f850e3034" Mar 13 08:54:39 crc kubenswrapper[4876]: I0313 08:54:39.036957 4876 scope.go:117] "RemoveContainer" containerID="4159d0a47a900f360cf9a2c6604a6e5a39e3dedf37546a271f3f2a0f11df17d1" Mar 13 08:54:39 crc kubenswrapper[4876]: E0313 08:54:39.038102 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r9cl2_openshift-machine-config-operator(0a6f71e5-2091-4386-b559-bba70bc45972)\"" pod="openshift-machine-config-operator/machine-config-daemon-r9cl2" podUID="0a6f71e5-2091-4386-b559-bba70bc45972" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515154750342024453 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015154750343017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015154737115016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015154737115015466 5ustar corecore